var/home/core/zuul-output/0000755000175000017500000000000015111060345014521 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111071123015461 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004755565515111071115017714 0ustar rootrootNov 24 13:46:23 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 13:46:23 crc restorecon[4689]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:23 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 13:46:24 crc restorecon[4689]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 13:46:25 crc kubenswrapper[4970]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.196605 4970 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202227 4970 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202263 4970 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202274 4970 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202284 4970 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202294 4970 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202303 4970 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202311 4970 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202320 4970 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202328 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202336 4970 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202344 4970 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202351 4970 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202359 4970 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202367 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202374 4970 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202382 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202390 4970 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202397 4970 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202406 4970 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202416 4970 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202427 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202435 4970 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202443 4970 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202452 4970 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202463 4970 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202478 4970 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202489 4970 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202498 4970 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202507 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202516 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202524 4970 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202533 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202541 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202551 4970 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202560 4970 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202568 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202605 4970 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202613 4970 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202621 4970 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202629 4970 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202637 4970 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202644 4970 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202652 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202660 4970 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202668 4970 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202676 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202683 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202691 4970 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202698 4970 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202706 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202714 4970 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202722 4970 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202732 4970 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202741 4970 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202749 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202758 4970 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202766 4970 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202774 4970 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202782 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202790 4970 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202798 4970 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202806 4970 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202815 4970 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202824 4970 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202834 4970 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202844 4970 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202854 4970 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202863 4970 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202876 4970 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202886 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.202897 4970 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203071 4970 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203088 4970 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203104 4970 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203115 4970 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203126 4970 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203136 4970 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203148 4970 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203159 4970 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203169 4970 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203178 4970 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203188 4970 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203197 4970 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203207 4970 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203216 4970 flags.go:64] FLAG: --cgroup-root="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203225 4970 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203237 4970 flags.go:64] FLAG: --client-ca-file="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203286 4970 flags.go:64] FLAG: --cloud-config="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203295 4970 flags.go:64] FLAG: --cloud-provider="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203305 4970 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203316 4970 flags.go:64] FLAG: --cluster-domain="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203326 4970 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203336 4970 flags.go:64] FLAG: --config-dir="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203344 4970 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203354 4970 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203375 4970 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203384 4970 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203394 4970 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203403 4970 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203412 4970 flags.go:64] FLAG: --contention-profiling="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203421 4970 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203430 4970 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203440 4970 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203449 4970 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203460 4970 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203469 4970 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203478 4970 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203519 4970 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203529 4970 flags.go:64] FLAG: --enable-server="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203537 4970 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203550 4970 flags.go:64] FLAG: --event-burst="100" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203560 4970 flags.go:64] FLAG: --event-qps="50" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203569 4970 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203605 4970 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203615 4970 flags.go:64] FLAG: --eviction-hard="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203627 4970 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203636 4970 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203646 4970 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203656 4970 flags.go:64] FLAG: --eviction-soft="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203665 4970 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203674 4970 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203683 4970 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203692 4970 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203702 4970 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203711 4970 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203720 4970 flags.go:64] FLAG: --feature-gates="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203730 4970 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203739 4970 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203749 4970 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203758 4970 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203768 4970 flags.go:64] FLAG: --healthz-port="10248" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203777 4970 flags.go:64] FLAG: --help="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203785 4970 flags.go:64] FLAG: --hostname-override="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203794 4970 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203804 4970 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203812 4970 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203821 4970 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203829 4970 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203840 4970 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203851 4970 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203862 4970 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203873 4970 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203885 4970 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203897 4970 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203909 4970 flags.go:64] FLAG: --kube-reserved="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203919 4970 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203928 4970 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203937 4970 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203946 4970 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203955 4970 flags.go:64] FLAG: --lock-file="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203972 4970 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203982 4970 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.203991 4970 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204004 4970 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204014 4970 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204023 4970 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204034 4970 flags.go:64] FLAG: --logging-format="text" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204045 4970 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204058 4970 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204070 4970 flags.go:64] FLAG: --manifest-url="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204081 4970 flags.go:64] FLAG: --manifest-url-header="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204098 4970 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204110 4970 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204124 4970 flags.go:64] FLAG: --max-pods="110" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204137 4970 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204149 4970 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204161 4970 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204173 4970 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204185 4970 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204196 4970 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204206 4970 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204227 4970 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204236 4970 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204245 4970 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204254 4970 flags.go:64] FLAG: --pod-cidr="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204263 4970 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204276 4970 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204285 4970 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204294 4970 flags.go:64] FLAG: --pods-per-core="0" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204303 4970 flags.go:64] FLAG: --port="10250" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204312 4970 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204321 4970 flags.go:64] FLAG: --provider-id="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204330 4970 flags.go:64] FLAG: --qos-reserved="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204339 4970 flags.go:64] FLAG: --read-only-port="10255" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204348 4970 flags.go:64] FLAG: --register-node="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204357 4970 flags.go:64] FLAG: --register-schedulable="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204367 4970 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204384 4970 flags.go:64] FLAG: --registry-burst="10" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204393 4970 flags.go:64] FLAG: --registry-qps="5" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204402 4970 flags.go:64] FLAG: --reserved-cpus="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204411 4970 flags.go:64] FLAG: --reserved-memory="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204444 4970 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204453 4970 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204463 4970 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204471 4970 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204481 4970 flags.go:64] FLAG: --runonce="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204489 4970 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204498 4970 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204508 4970 flags.go:64] FLAG: --seccomp-default="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204517 4970 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204527 4970 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204536 4970 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204545 4970 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204554 4970 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204563 4970 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204613 4970 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204625 4970 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204634 4970 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204643 4970 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204652 4970 flags.go:64] FLAG: --system-cgroups="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204661 4970 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204675 4970 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204684 4970 flags.go:64] FLAG: --tls-cert-file="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204694 4970 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204704 4970 flags.go:64] FLAG: --tls-min-version="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204714 4970 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204724 4970 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204734 4970 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204744 4970 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204753 4970 flags.go:64] FLAG: --v="2" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204764 4970 flags.go:64] FLAG: --version="false" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204776 4970 flags.go:64] FLAG: --vmodule="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204787 4970 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.204797 4970 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205043 4970 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205056 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205065 4970 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205073 4970 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205081 4970 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205089 4970 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205097 4970 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205104 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205113 4970 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205123 4970 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205132 4970 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205141 4970 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205151 4970 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205162 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205171 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205179 4970 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205188 4970 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205196 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205204 4970 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205214 4970 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205224 4970 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205235 4970 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205245 4970 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205256 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205269 4970 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205279 4970 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205291 4970 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205301 4970 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205313 4970 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205324 4970 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205335 4970 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205346 4970 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205356 4970 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205366 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205380 4970 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205391 4970 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205401 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205409 4970 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205418 4970 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205426 4970 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205434 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205470 4970 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205481 4970 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205492 4970 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205500 4970 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205509 4970 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205519 4970 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205530 4970 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205538 4970 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205546 4970 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205556 4970 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205564 4970 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205605 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205615 4970 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205623 4970 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205640 4970 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205649 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205657 4970 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205665 4970 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205673 4970 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205681 4970 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205690 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205697 4970 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205705 4970 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205713 4970 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205721 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205730 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205737 4970 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205745 4970 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205753 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.205762 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.205784 4970 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.217414 4970 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.217444 4970 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217526 4970 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217534 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217540 4970 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217545 4970 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217550 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217557 4970 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217563 4970 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217568 4970 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217589 4970 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217594 4970 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217599 4970 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217605 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217611 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217618 4970 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217625 4970 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217631 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217637 4970 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217643 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217648 4970 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217652 4970 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217659 4970 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217667 4970 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217675 4970 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217681 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217686 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217691 4970 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217696 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217701 4970 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217706 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217711 4970 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217716 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217721 4970 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217726 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217731 4970 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217737 4970 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217742 4970 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217746 4970 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217750 4970 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217755 4970 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217760 4970 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217765 4970 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217770 4970 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217774 4970 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217779 4970 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217784 4970 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217789 4970 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217794 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217799 4970 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217804 4970 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217809 4970 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217813 4970 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217818 4970 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217822 4970 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217829 4970 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217834 4970 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217838 4970 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217845 4970 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217851 4970 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217856 4970 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217862 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217867 4970 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217871 4970 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217877 4970 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217881 4970 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217886 4970 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217891 4970 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217895 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217900 4970 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217906 4970 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217912 4970 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.217918 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.217926 4970 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218061 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218074 4970 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218079 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218084 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218088 4970 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218093 4970 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218097 4970 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218103 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218109 4970 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218118 4970 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218124 4970 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218129 4970 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218134 4970 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218140 4970 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218145 4970 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218149 4970 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218154 4970 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218159 4970 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218163 4970 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218168 4970 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218172 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218177 4970 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218181 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218186 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218190 4970 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218194 4970 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218199 4970 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218203 4970 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218208 4970 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218212 4970 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218217 4970 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218221 4970 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218226 4970 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218230 4970 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218235 4970 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218240 4970 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218245 4970 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218249 4970 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218254 4970 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218259 4970 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218263 4970 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218269 4970 feature_gate.go:330] unrecognized feature gate: Example Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218274 4970 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218278 4970 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218283 4970 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218287 4970 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218292 4970 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218297 4970 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218302 4970 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218308 4970 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218313 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218318 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218323 4970 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218328 4970 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218333 4970 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218337 4970 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218342 4970 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218349 4970 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218356 4970 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218362 4970 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218368 4970 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218375 4970 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218380 4970 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218387 4970 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218392 4970 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218397 4970 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218402 4970 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218406 4970 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218411 4970 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218415 4970 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.218421 4970 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.218428 4970 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.218594 4970 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.224203 4970 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.224291 4970 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.226134 4970 server.go:997] "Starting client certificate rotation" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.226166 4970 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.226326 4970 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-20 05:00:33.640966357 +0000 UTC Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.226394 4970 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 615h14m8.414577046s for next certificate rotation Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.267325 4970 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.269352 4970 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.297202 4970 log.go:25] "Validated CRI v1 runtime API" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.341058 4970 log.go:25] "Validated CRI v1 image API" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.343644 4970 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.349877 4970 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-13-41-59-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.349926 4970 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.378525 4970 manager.go:217] Machine: {Timestamp:2025-11-24 13:46:25.3748035 +0000 UTC m=+0.662560873 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590 BootID:40be1732-2104-461a-99c4-0be47e2bfcda Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:25:d4:c2 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:25:d4:c2 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a2:a4:7e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e3:79:20 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:40:85:d5 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:9d:7f:e6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:de:62:3e:0f:45:40 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f2:09:b1:00:95:44 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.378933 4970 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.379110 4970 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.381526 4970 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.381965 4970 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.382025 4970 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.382332 4970 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.382352 4970 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.382768 4970 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.382818 4970 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.383036 4970 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.383170 4970 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.388361 4970 kubelet.go:418] "Attempting to sync node with API server" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.388616 4970 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.388700 4970 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.388728 4970 kubelet.go:324] "Adding apiserver pod source" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.388746 4970 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.393174 4970 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.394541 4970 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.397771 4970 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.398450 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.398501 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.398569 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.398654 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.399970 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400021 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400036 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400049 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400070 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400084 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400098 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400120 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400134 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400150 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400197 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.400212 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.402489 4970 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.403141 4970 server.go:1280] "Started kubelet" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.407096 4970 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.407420 4970 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 13:46:25 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.409337 4970 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.410658 4970 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.410744 4970 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.410871 4970 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 04:30:52.67944011 +0000 UTC Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.410909 4970 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 686h44m27.268534802s for next certificate rotation Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.410904 4970 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.411022 4970 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.411068 4970 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.411122 4970 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.411620 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="200ms" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.412701 4970 factory.go:55] Registering systemd factory Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.412723 4970 factory.go:221] Registration of the systemd container factory successfully Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.418819 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.418924 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.412774 4970 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.419657 4970 factory.go:153] Registering CRI-O factory Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.419680 4970 factory.go:221] Registration of the crio container factory successfully Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.419747 4970 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.419769 4970 factory.go:103] Registering Raw factory Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.419787 4970 manager.go:1196] Started watching for new ooms in manager Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.420367 4970 manager.go:319] Starting recovery of all containers Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.421516 4970 server.go:460] "Adding debug handlers to kubelet server" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.419232 4970 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.12:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187af55d5c967c4e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 13:46:25.403100238 +0000 UTC m=+0.690857561,LastTimestamp:2025-11-24 13:46:25.403100238 +0000 UTC m=+0.690857561,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431266 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431309 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431323 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431341 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431404 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431422 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431435 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431449 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431469 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.431481 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432119 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432335 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432408 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432466 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432513 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432544 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432615 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432645 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432666 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432696 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432717 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432745 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432766 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432787 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432822 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432850 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432898 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.432929 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433095 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433127 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433165 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433192 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433226 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433254 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433282 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433318 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433339 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433359 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433389 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433409 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433438 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433461 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433481 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433508 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433529 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433558 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433607 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433639 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433667 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433687 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433712 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433731 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433766 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433789 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433816 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433845 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433871 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433899 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433919 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433938 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433965 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.433988 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434016 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434036 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434055 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434202 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434251 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434300 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434334 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434365 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434409 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434438 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434478 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434509 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434540 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434610 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434643 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434678 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434704 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434734 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434769 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434794 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434832 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434861 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434889 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.434924 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437432 4970 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437518 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437556 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437654 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437689 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437715 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437741 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437764 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437789 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437826 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437875 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437913 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437942 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.437971 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438062 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438099 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438138 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438169 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438199 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438252 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438300 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438347 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438485 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438530 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438565 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438636 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438681 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438746 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438803 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438815 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438830 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438840 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.438850 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440532 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440838 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440860 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440876 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440892 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440906 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440920 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440934 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440950 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440964 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440977 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.440991 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441005 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441019 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441032 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441048 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441064 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441076 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441089 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441102 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441116 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441130 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441143 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441156 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441170 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441185 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441201 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441216 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441230 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441243 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441257 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441271 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441284 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441299 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441316 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441339 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441352 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441365 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441379 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441393 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441406 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441418 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441431 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441445 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441463 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441481 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441498 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441511 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441526 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441539 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441553 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441567 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441604 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441618 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441630 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441644 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441657 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441670 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441684 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441698 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441711 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441725 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441738 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441752 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441765 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441779 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441793 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441809 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441824 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441838 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441852 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441866 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441879 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441892 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441906 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441919 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441932 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441950 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441963 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441977 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.441991 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442004 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442018 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442031 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442043 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442057 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442070 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442084 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442098 4970 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442112 4970 reconstruct.go:97] "Volume reconstruction finished" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.442121 4970 reconciler.go:26] "Reconciler: start to sync state" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.451541 4970 manager.go:324] Recovery completed Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.463910 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.465009 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.465140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.465214 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.466169 4970 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.466260 4970 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.466343 4970 state_mem.go:36] "Initialized new in-memory state store" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.466587 4970 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.469051 4970 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.469092 4970 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.469116 4970 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.469191 4970 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.471474 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.471566 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.486635 4970 policy_none.go:49] "None policy: Start" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.489146 4970 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.489250 4970 state_mem.go:35] "Initializing new in-memory state store" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.512029 4970 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.541668 4970 manager.go:334] "Starting Device Plugin manager" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.541721 4970 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.541742 4970 server.go:79] "Starting device plugin registration server" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.542223 4970 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.542241 4970 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.542432 4970 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.542526 4970 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.542537 4970 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.550746 4970 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.569955 4970 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.570026 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572261 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572293 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572304 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572416 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572654 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.572690 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573175 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573257 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573274 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573285 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573294 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573554 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573638 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573938 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.573978 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.574069 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.574187 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.574217 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575025 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575321 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575345 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575419 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.575439 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576179 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576534 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.576634 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.577572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.577640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.577658 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.577899 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.577941 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.578947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.578966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.578974 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.613272 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="400ms" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.642410 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643102 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643175 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643230 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643260 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643384 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643458 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643499 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643543 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643609 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643639 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643671 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643821 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643871 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643916 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.643958 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.644270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.644320 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.644345 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.644390 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.645013 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745748 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745836 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745880 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745915 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745926 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745992 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746019 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745945 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.745992 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746024 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746133 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746086 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746215 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746250 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746272 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746298 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746311 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746335 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746319 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746364 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746422 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746453 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746464 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746497 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746501 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746558 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746572 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746683 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746700 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.746736 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.845989 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.848140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.848223 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.848243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.848279 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:25 crc kubenswrapper[4970]: E1124 13:46:25.848890 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.913948 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.939651 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.948354 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.956789 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-62e469b672c9eba4a8c2a469119d197d3f8770f2ba3b27e5f7ee6bf1ccd0baeb WatchSource:0}: Error finding container 62e469b672c9eba4a8c2a469119d197d3f8770f2ba3b27e5f7ee6bf1ccd0baeb: Status 404 returned error can't find the container with id 62e469b672c9eba4a8c2a469119d197d3f8770f2ba3b27e5f7ee6bf1ccd0baeb Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.966327 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.972199 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-61abaaba2bfb8e46b5a7c0493190a78bc91f54cdca4ff06a59e9911b5ddb0e41 WatchSource:0}: Error finding container 61abaaba2bfb8e46b5a7c0493190a78bc91f54cdca4ff06a59e9911b5ddb0e41: Status 404 returned error can't find the container with id 61abaaba2bfb8e46b5a7c0493190a78bc91f54cdca4ff06a59e9911b5ddb0e41 Nov 24 13:46:25 crc kubenswrapper[4970]: I1124 13:46:25.973741 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.975805 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ade59a7dca22b807660e5a143455cfc9fdecbb45bfbbf01f059608885617dbe5 WatchSource:0}: Error finding container ade59a7dca22b807660e5a143455cfc9fdecbb45bfbbf01f059608885617dbe5: Status 404 returned error can't find the container with id ade59a7dca22b807660e5a143455cfc9fdecbb45bfbbf01f059608885617dbe5 Nov 24 13:46:25 crc kubenswrapper[4970]: W1124 13:46:25.990905 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e9810dd083fdc9dbf57fd651c619a8e72f067aea917e14fc1f787eaf8a52f22c WatchSource:0}: Error finding container e9810dd083fdc9dbf57fd651c619a8e72f067aea917e14fc1f787eaf8a52f22c: Status 404 returned error can't find the container with id e9810dd083fdc9dbf57fd651c619a8e72f067aea917e14fc1f787eaf8a52f22c Nov 24 13:46:26 crc kubenswrapper[4970]: W1124 13:46:26.000448 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ce1fa51bf8589c6a4fd0f56189ab302bc163687aca86f5157eced00fb470b799 WatchSource:0}: Error finding container ce1fa51bf8589c6a4fd0f56189ab302bc163687aca86f5157eced00fb470b799: Status 404 returned error can't find the container with id ce1fa51bf8589c6a4fd0f56189ab302bc163687aca86f5157eced00fb470b799 Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.014249 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="800ms" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.249564 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.251439 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.251489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.251502 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.251530 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.252011 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 24 13:46:26 crc kubenswrapper[4970]: W1124 13:46:26.337925 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.338033 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:26 crc kubenswrapper[4970]: W1124 13:46:26.380472 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.380629 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:26 crc kubenswrapper[4970]: W1124 13:46:26.398089 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.398168 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.420316 4970 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.472439 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"62e469b672c9eba4a8c2a469119d197d3f8770f2ba3b27e5f7ee6bf1ccd0baeb"} Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.473331 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ce1fa51bf8589c6a4fd0f56189ab302bc163687aca86f5157eced00fb470b799"} Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.474147 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e9810dd083fdc9dbf57fd651c619a8e72f067aea917e14fc1f787eaf8a52f22c"} Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.475148 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ade59a7dca22b807660e5a143455cfc9fdecbb45bfbbf01f059608885617dbe5"} Nov 24 13:46:26 crc kubenswrapper[4970]: I1124 13:46:26.476050 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"61abaaba2bfb8e46b5a7c0493190a78bc91f54cdca4ff06a59e9911b5ddb0e41"} Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.815687 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="1.6s" Nov 24 13:46:26 crc kubenswrapper[4970]: W1124 13:46:26.836380 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:26 crc kubenswrapper[4970]: E1124 13:46:26.836449 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.052966 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.055261 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.055315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.055333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.055365 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:27 crc kubenswrapper[4970]: E1124 13:46:27.055933 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.420222 4970 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.480465 4970 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8" exitCode=0 Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.480535 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.480569 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.481853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.481903 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.481923 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.482447 4970 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf" exitCode=0 Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.482516 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.482673 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.483758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.483798 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.483813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.483986 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.484476 4970 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5c00be1e235f219adb7db84622e42ac755e988a852008077f9616b24a5439746" exitCode=0 Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.484533 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5c00be1e235f219adb7db84622e42ac755e988a852008077f9616b24a5439746"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.484734 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.486814 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.486867 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.486885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.486829 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.487041 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.487059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.489183 4970 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f" exitCode=0 Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.489287 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.489310 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.490625 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.490672 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.490690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.497653 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.497740 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.497769 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.497794 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422"} Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.498040 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.499364 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.499421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:27 crc kubenswrapper[4970]: I1124 13:46:27.499439 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: W1124 13:46:28.067025 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:28 crc kubenswrapper[4970]: E1124 13:46:28.067094 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:28 crc kubenswrapper[4970]: W1124 13:46:28.241783 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:28 crc kubenswrapper[4970]: E1124 13:46:28.241883 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:28 crc kubenswrapper[4970]: E1124 13:46:28.416953 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.12:6443: connect: connection refused" interval="3.2s" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.419885 4970 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.503010 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.503054 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.503064 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.503142 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.504777 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.504837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.504862 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.507418 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.507476 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.507503 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.507529 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.509642 4970 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29" exitCode=0 Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.509736 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.509879 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.510885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.510959 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.510985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.514715 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dbde4abe7fe171ed1ae8e8dbb3ff7e82ba95e8bf97c343ad12ffec98285998bb"} Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.515052 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.515305 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.520918 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.520983 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.521032 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.524463 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.524526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.524615 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.658195 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.659836 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.659893 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.659910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.659941 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:28 crc kubenswrapper[4970]: E1124 13:46:28.660513 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.12:6443: connect: connection refused" node="crc" Nov 24 13:46:28 crc kubenswrapper[4970]: W1124 13:46:28.749481 4970 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.12:6443: connect: connection refused Nov 24 13:46:28 crc kubenswrapper[4970]: E1124 13:46:28.749567 4970 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.12:6443: connect: connection refused" logger="UnhandledError" Nov 24 13:46:28 crc kubenswrapper[4970]: I1124 13:46:28.782768 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.177334 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.522355 4970 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749" exitCode=0 Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.522439 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749"} Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.522629 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.524894 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.525076 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.525204 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.528628 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.529359 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.529497 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.529818 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea"} Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.530126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.530173 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.530197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.530494 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.531958 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.531971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532029 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532748 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.532939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:29 crc kubenswrapper[4970]: I1124 13:46:29.533079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.537842 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a"} Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.537961 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8"} Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.537993 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc"} Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.538004 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.537909 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.538146 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.539527 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.539607 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.539633 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.540102 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.540182 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.540221 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.708500 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.708742 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.710425 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.710494 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.710513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:30 crc kubenswrapper[4970]: I1124 13:46:30.716929 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.550213 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1"} Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.550284 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e"} Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.550314 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.551008 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.551691 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.551733 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.551746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.552370 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.552414 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.552431 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.783365 4970 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.783463 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.861282 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.863234 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.863292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.863310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:31 crc kubenswrapper[4970]: I1124 13:46:31.863343 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.370637 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.552693 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.552699 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554219 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554266 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554305 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554279 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:32 crc kubenswrapper[4970]: I1124 13:46:32.554326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.657400 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.657540 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.657603 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.659073 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.659143 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:33 crc kubenswrapper[4970]: I1124 13:46:33.659155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.012275 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.114409 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.114690 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.116070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.116171 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.116190 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.559249 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.559330 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.560741 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.560802 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.560828 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.593065 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.593266 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.594295 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.594355 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:34 crc kubenswrapper[4970]: I1124 13:46:34.594381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:35 crc kubenswrapper[4970]: I1124 13:46:35.143631 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:35 crc kubenswrapper[4970]: E1124 13:46:35.551016 4970 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 13:46:35 crc kubenswrapper[4970]: I1124 13:46:35.562052 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:35 crc kubenswrapper[4970]: I1124 13:46:35.563572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:35 crc kubenswrapper[4970]: I1124 13:46:35.563690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:35 crc kubenswrapper[4970]: I1124 13:46:35.563885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.151761 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.152028 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.153758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.153819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.153838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.158665 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.565474 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.566499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.566610 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:36 crc kubenswrapper[4970]: I1124 13:46:36.566635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:39 crc kubenswrapper[4970]: I1124 13:46:39.417938 4970 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 13:46:39 crc kubenswrapper[4970]: I1124 13:46:39.418898 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:46:39 crc kubenswrapper[4970]: I1124 13:46:39.437638 4970 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 13:46:39 crc kubenswrapper[4970]: I1124 13:46:39.437733 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 13:46:41 crc kubenswrapper[4970]: I1124 13:46:41.784722 4970 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 13:46:41 crc kubenswrapper[4970]: I1124 13:46:41.784830 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.662907 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.663079 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.664159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.664209 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.664228 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:43 crc kubenswrapper[4970]: I1124 13:46:43.668774 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.381834 4970 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.385876 4970 trace.go:236] Trace[1800560111]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:46:32.766) (total time: 11618ms): Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1800560111]: ---"Objects listed" error: 11618ms (13:46:44.385) Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1800560111]: [11.618996888s] [11.618996888s] END Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.385943 4970 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.386177 4970 trace.go:236] Trace[1016316663]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:46:29.499) (total time: 14886ms): Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1016316663]: ---"Objects listed" error: 14886ms (13:46:44.386) Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1016316663]: [14.886906733s] [14.886906733s] END Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.386205 4970 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.386602 4970 trace.go:236] Trace[1169186496]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:46:32.938) (total time: 11448ms): Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1169186496]: ---"Objects listed" error: 11448ms (13:46:44.386) Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1169186496]: [11.448152505s] [11.448152505s] END Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.386633 4970 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.392852 4970 trace.go:236] Trace[1845708820]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 13:46:31.893) (total time: 12499ms): Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1845708820]: ---"Objects listed" error: 12499ms (13:46:44.392) Nov 24 13:46:44 crc kubenswrapper[4970]: Trace[1845708820]: [12.499416521s] [12.499416521s] END Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.392885 4970 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.394831 4970 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.395262 4970 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.403779 4970 apiserver.go:52] "Watching apiserver" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.408983 4970 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.409236 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.409748 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.409800 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.409839 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.409869 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.410150 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.410173 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.410234 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.410426 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.410565 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.411798 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.412649 4970 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.412777 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.412821 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.413006 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.418543 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.421424 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.421521 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.421812 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.422082 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.468930 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.481153 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.491606 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496074 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496128 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496152 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496174 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496194 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496215 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496235 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496256 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496278 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496297 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496349 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496400 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496421 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496443 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496463 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496487 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496510 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496534 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496530 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496556 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496598 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496628 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496650 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496653 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496670 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496653 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496693 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496736 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496758 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496808 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496847 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496898 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496903 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496924 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496942 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496948 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496973 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496972 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.496986 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497035 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497062 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497091 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497116 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497142 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497153 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497163 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497186 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497208 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497238 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497265 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497289 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497312 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497334 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497356 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497378 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497398 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497423 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497443 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497466 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497487 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497547 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497613 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497639 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497660 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497680 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497700 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497722 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497743 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497766 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497758 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497788 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497812 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497825 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497834 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497885 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497906 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497928 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497951 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.497999 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498019 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498044 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498070 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498130 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498152 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498173 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498193 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498219 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498248 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498273 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498293 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498313 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498334 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498354 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498360 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498374 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498423 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498463 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498487 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498511 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498532 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498552 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498589 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498610 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498630 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498654 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498675 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498698 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498719 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498741 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498762 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498782 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498803 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498825 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498847 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498867 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498887 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498911 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498931 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498950 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498973 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.498994 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499013 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499035 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499059 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499080 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499100 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499761 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499787 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499807 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499828 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499849 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499872 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499896 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499918 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499940 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499963 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499988 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500011 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500034 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500056 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500081 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500102 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500124 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500148 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500169 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500189 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500210 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500231 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500309 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500333 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500365 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500398 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500431 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500470 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500492 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500516 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500538 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500560 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500623 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500646 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500668 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500689 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500712 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500734 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500754 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500776 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500802 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500823 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500846 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500867 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500890 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500916 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500986 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501013 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501036 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501060 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501087 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501110 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501195 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501222 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501247 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501270 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501292 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501315 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501337 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501360 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501384 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501406 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501429 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501451 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501472 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501494 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501517 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501540 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501564 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501628 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501650 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501673 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501697 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501719 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501744 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501768 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501792 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501816 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501838 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501871 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501905 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501934 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501976 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502009 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502072 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502114 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502152 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502192 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502230 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502265 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502294 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502330 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502364 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502397 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502433 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502469 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502504 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502541 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502640 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502661 4970 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502680 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502699 4970 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502718 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502735 4970 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502753 4970 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502772 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502793 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502815 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502833 4970 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502852 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502872 4970 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502892 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502912 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.509097 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.513327 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.514135 4970 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.520455 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530709 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499701 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.499907 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500892 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500959 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.500884 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501103 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501130 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501217 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501253 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501536 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.501779 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502109 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502228 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502309 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502844 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502640 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.502892 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.503119 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.503243 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.503378 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.503485 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504148 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504203 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504383 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504680 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504944 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504965 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.504968 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.505003 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.505099 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.505872 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.505946 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.506076 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.506160 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.506289 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.506341 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.506924 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.508842 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.509440 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.509548 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.509746 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.510687 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.510874 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.510998 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.511123 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.511468 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.511563 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.511617 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531778 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.511794 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.512884 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.515807 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.515815 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.516212 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.516787 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.516802 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.517062 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.517916 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.517990 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.518211 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.518687 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.518761 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.518776 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.518917 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.519964 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.520861 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.520941 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:46:45.020922406 +0000 UTC m=+20.308679709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.521340 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.521682 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.521995 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.522510 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.523211 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.523696 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.524036 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.524112 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.524117 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.524210 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.524509 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.525052 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.525887 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.526262 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.526426 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.527518 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.528291 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.528676 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.528795 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.529348 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.529667 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530041 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530097 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530325 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530396 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530459 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530705 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.530757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531021 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531313 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531332 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531478 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.531671 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.532373 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.532548 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.533403 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.533465 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.533836 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.534081 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.534267 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.534493 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.534775 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.535355 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.535555 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.535589 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.535603 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.535743 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.535757 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.535794 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.535915 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.536048 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:45.036009676 +0000 UTC m=+20.323767019 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.536090 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:45.036074487 +0000 UTC m=+20.323831820 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.536143 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:45.036132899 +0000 UTC m=+20.323890232 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:44 crc kubenswrapper[4970]: E1124 13:46:44.536167 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:45.036156619 +0000 UTC m=+20.323913952 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.536211 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.536457 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.536555 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.536760 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.537219 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.537709 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.537878 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.539755 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.539805 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.539813 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.540638 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541218 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.537019 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541727 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541812 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.537685 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541918 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541954 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.541968 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542154 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542169 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542273 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542321 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542481 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542555 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542892 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542950 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.542992 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.543434 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.543679 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.543745 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.543884 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.545720 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.547407 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.548535 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.554124 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.554438 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.554612 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.556234 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.556902 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.557392 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558131 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558200 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558261 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558384 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558517 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.558545 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.559491 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.560043 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.560080 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.554195 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.560878 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.561032 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.561737 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.561747 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.563621 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.566769 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.566870 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.567117 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.567500 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.567904 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.568003 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.568483 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.568774 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.568839 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.573480 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.574695 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.575488 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.576790 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.577115 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.577178 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.577367 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.577526 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.578150 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.578261 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.586085 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.587764 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.587978 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.600384 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603698 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603742 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603786 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603797 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603806 4970 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603814 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603823 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603833 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603841 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603849 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603857 4970 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603865 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603873 4970 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603882 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603890 4970 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603897 4970 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603897 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603936 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603905 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603970 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603979 4970 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603988 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.603997 4970 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604007 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604015 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604023 4970 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604031 4970 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604040 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604048 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604056 4970 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604064 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604071 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604080 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604088 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604096 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604104 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604113 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604121 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604130 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604139 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604147 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604154 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604162 4970 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604170 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604179 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604186 4970 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604195 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604204 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604212 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604220 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604227 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604235 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604243 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604253 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604261 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604269 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604278 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604287 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604295 4970 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604303 4970 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604311 4970 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604321 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604330 4970 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604338 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604346 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604355 4970 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604362 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604371 4970 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604379 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604388 4970 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604396 4970 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604404 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604412 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604420 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604428 4970 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604437 4970 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604445 4970 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604453 4970 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604461 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604469 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604477 4970 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604485 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604493 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604509 4970 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604518 4970 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604526 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604534 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604542 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604550 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604558 4970 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604566 4970 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604589 4970 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604598 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604606 4970 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604615 4970 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604623 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604631 4970 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604639 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604647 4970 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604655 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604672 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604680 4970 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604689 4970 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604697 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604704 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604712 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604720 4970 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604728 4970 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604735 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604743 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604751 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604759 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604768 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604777 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604810 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604818 4970 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604831 4970 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604838 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604846 4970 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604853 4970 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604861 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604870 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604878 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604886 4970 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604894 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604901 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604909 4970 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604916 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604924 4970 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604931 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604939 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604947 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604955 4970 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604979 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604988 4970 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.604995 4970 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605003 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605011 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605019 4970 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605026 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605033 4970 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605041 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605048 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605056 4970 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605064 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605071 4970 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605078 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605086 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605093 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605101 4970 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605108 4970 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605116 4970 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605123 4970 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605131 4970 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605139 4970 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605146 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605154 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605161 4970 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605168 4970 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605176 4970 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605183 4970 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605190 4970 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605199 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605207 4970 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605215 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605225 4970 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605233 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605241 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605248 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605256 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605284 4970 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605291 4970 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605299 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605306 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605314 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605322 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605329 4970 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605336 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605344 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605351 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.605359 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.606857 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.607757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.624706 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.633703 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.637270 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.644805 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.651741 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.653430 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.661570 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.678807 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.705780 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.705804 4970 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.708939 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.723621 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.732301 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.740667 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.741680 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.750831 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.752037 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.752133 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: W1124 13:46:44.752692 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-cc7a6a5210fe7e8b0a3ff9c2821ac0f30f695fb16a4eca68d197cd6953faaa41 WatchSource:0}: Error finding container cc7a6a5210fe7e8b0a3ff9c2821ac0f30f695fb16a4eca68d197cd6953faaa41: Status 404 returned error can't find the container with id cc7a6a5210fe7e8b0a3ff9c2821ac0f30f695fb16a4eca68d197cd6953faaa41 Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.759398 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.762237 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.770691 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: W1124 13:46:44.774791 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-0c5a10161c607f09eb566096aeac8fc38fb5bdeade9b02face2c47ea197456fd WatchSource:0}: Error finding container 0c5a10161c607f09eb566096aeac8fc38fb5bdeade9b02face2c47ea197456fd: Status 404 returned error can't find the container with id 0c5a10161c607f09eb566096aeac8fc38fb5bdeade9b02face2c47ea197456fd Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.781627 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.836557 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.846553 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.855777 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.864265 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.884034 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.905839 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.926116 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.943421 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:44 crc kubenswrapper[4970]: I1124 13:46:44.955967 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.109706 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.109809 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:46:46.109773914 +0000 UTC m=+21.397531207 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.109854 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.109922 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.109960 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.109981 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110051 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110068 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110070 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110081 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110094 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110091 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110106 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110093 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:46.110084261 +0000 UTC m=+21.397841554 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110252 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:46.110226955 +0000 UTC m=+21.397984318 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110271 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:46.110262806 +0000 UTC m=+21.398020119 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110093 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.110315 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:46.110304947 +0000 UTC m=+21.398062360 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.470002 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.470267 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.478048 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.479228 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.481662 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.483101 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.484360 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.485452 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.488310 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.489687 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.491883 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.492940 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.495002 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.496694 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.496671 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.498468 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.499434 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.500726 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.501386 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.502230 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.503312 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.504091 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.504837 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.505962 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.506762 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.507855 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.508757 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.509311 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.511070 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.512750 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.513953 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.515405 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.516671 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.517756 4970 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.517964 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.522802 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.523167 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.524539 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.525708 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.527979 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.529255 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.529925 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.531285 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.532249 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.533700 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.535614 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.537972 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.540203 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.541270 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.542472 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.544460 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.547609 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.549153 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.550719 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.553276 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.554358 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.555409 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.557670 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.559623 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.580665 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.592162 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.592332 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.592443 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0d8a832bc114cadadcaa58f32b35924a0f14d8c6d0d7d43e3463a3dd7952d5c1"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.593850 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.593903 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.599064 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cc7a6a5210fe7e8b0a3ff9c2821ac0f30f695fb16a4eca68d197cd6953faaa41"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.600484 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0c5a10161c607f09eb566096aeac8fc38fb5bdeade9b02face2c47ea197456fd"} Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.607557 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: E1124 13:46:45.608643 4970 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.624337 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.638064 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.651176 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.662268 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.676515 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.692024 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.706037 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.719778 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.738986 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:45 crc kubenswrapper[4970]: I1124 13:46:45.756875 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.118637 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.118754 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.118814 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:46:48.118786889 +0000 UTC m=+23.406544182 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.118863 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.118900 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.118904 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.118929 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.118956 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:48.118944453 +0000 UTC m=+23.406701746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119016 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119027 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119038 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119073 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:48.119058916 +0000 UTC m=+23.406816209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119113 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119170 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119198 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119218 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:48.119193089 +0000 UTC m=+23.406950412 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119219 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.119284 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:48.119266531 +0000 UTC m=+23.407023864 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.470222 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:46 crc kubenswrapper[4970]: I1124 13:46:46.470314 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.470364 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:46 crc kubenswrapper[4970]: E1124 13:46:46.470450 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:47 crc kubenswrapper[4970]: I1124 13:46:47.470289 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:47 crc kubenswrapper[4970]: E1124 13:46:47.470564 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.137951 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.138035 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.138068 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.138094 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138160 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138185 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:46:52.138145482 +0000 UTC m=+27.425902805 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138220 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138235 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:52.138221024 +0000 UTC m=+27.425978347 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138352 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:52.138301196 +0000 UTC m=+27.426058529 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.138405 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138454 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138516 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138533 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138546 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138558 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138617 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138673 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:52.138659395 +0000 UTC m=+27.426416728 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.138698 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:46:52.138687046 +0000 UTC m=+27.426444379 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.470044 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.470045 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.470280 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:48 crc kubenswrapper[4970]: E1124 13:46:48.470394 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.610955 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae"} Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.632486 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.654271 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.672506 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.694674 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.709366 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.732703 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.745609 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.761001 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.787080 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.792917 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.796140 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.801500 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.817134 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.832884 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.845635 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.861295 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.880922 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.903911 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.926634 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.940469 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.953534 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.969487 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.980773 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:48 crc kubenswrapper[4970]: I1124 13:46:48.993807 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:49 crc kubenswrapper[4970]: I1124 13:46:49.006897 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:49 crc kubenswrapper[4970]: I1124 13:46:49.025930 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:49 crc kubenswrapper[4970]: I1124 13:46:49.039826 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:49 crc kubenswrapper[4970]: I1124 13:46:49.051341 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:49 crc kubenswrapper[4970]: I1124 13:46:49.470109 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:49 crc kubenswrapper[4970]: E1124 13:46:49.470240 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.074671 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bhjg8"] Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.075122 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-vtfnd"] Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.075345 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.075529 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.077641 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.077830 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.078296 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.078379 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.078653 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.078801 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.078977 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.080733 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.094233 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.112020 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.142800 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155281 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-os-release\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155331 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-k8s-cni-cncf-io\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155377 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-netns\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155405 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-hostroot\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155451 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn7wj\" (UniqueName: \"kubernetes.io/projected/0875a754-321a-4a3a-a798-941eefcdffe5-kube-api-access-cn7wj\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155478 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-cni-binary-copy\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155504 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-etc-kubernetes\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155530 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-system-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155557 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-conf-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155616 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-multus-daemon-config\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155653 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f2628d9f-c695-4d55-bf97-30c4443ca0d3-hosts-file\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155692 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-socket-dir-parent\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155734 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-multus\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155792 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-bin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155836 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmfvl\" (UniqueName: \"kubernetes.io/projected/f2628d9f-c695-4d55-bf97-30c4443ca0d3-kube-api-access-xmfvl\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155879 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-multus-certs\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155953 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.155997 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-kubelet\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.156043 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-cnibin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.160290 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.178071 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.189938 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.235042 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.256924 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-socket-dir-parent\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.256968 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-multus\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.256992 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-bin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257015 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmfvl\" (UniqueName: \"kubernetes.io/projected/f2628d9f-c695-4d55-bf97-30c4443ca0d3-kube-api-access-xmfvl\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257040 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257062 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-multus-certs\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257100 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-cnibin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257100 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-bin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257128 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-kubelet\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257151 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-cnibin\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257096 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-cni-multus\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257100 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-socket-dir-parent\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257153 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-os-release\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257201 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-var-lib-kubelet\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257273 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-k8s-cni-cncf-io\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257298 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-netns\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257306 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257320 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-k8s-cni-cncf-io\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257339 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-netns\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257317 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-hostroot\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257125 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-host-run-multus-certs\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257393 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-etc-kubernetes\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257340 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-hostroot\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257417 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn7wj\" (UniqueName: \"kubernetes.io/projected/0875a754-321a-4a3a-a798-941eefcdffe5-kube-api-access-cn7wj\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257452 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-etc-kubernetes\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257478 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-cni-binary-copy\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257510 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-system-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257518 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-os-release\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257532 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-conf-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257556 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-multus-daemon-config\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257588 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-multus-conf-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257593 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f2628d9f-c695-4d55-bf97-30c4443ca0d3-hosts-file\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257635 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f2628d9f-c695-4d55-bf97-30c4443ca0d3-hosts-file\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.257608 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0875a754-321a-4a3a-a798-941eefcdffe5-system-cni-dir\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.258269 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-multus-daemon-config\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.258282 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0875a754-321a-4a3a-a798-941eefcdffe5-cni-binary-copy\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.265498 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.286113 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmfvl\" (UniqueName: \"kubernetes.io/projected/f2628d9f-c695-4d55-bf97-30c4443ca0d3-kube-api-access-xmfvl\") pod \"node-resolver-bhjg8\" (UID: \"f2628d9f-c695-4d55-bf97-30c4443ca0d3\") " pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.286135 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn7wj\" (UniqueName: \"kubernetes.io/projected/0875a754-321a-4a3a-a798-941eefcdffe5-kube-api-access-cn7wj\") pod \"multus-vtfnd\" (UID: \"0875a754-321a-4a3a-a798-941eefcdffe5\") " pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.286793 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.294999 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.305759 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.316457 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.326819 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.335938 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.346325 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.363615 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.375763 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.386451 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.388600 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bhjg8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.395284 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.396303 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-vtfnd" Nov 24 13:46:50 crc kubenswrapper[4970]: W1124 13:46:50.398855 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2628d9f_c695_4d55_bf97_30c4443ca0d3.slice/crio-940ac92637132558cd794b90ed14e86c6d3c6fc9b5485452639b3a68c6c91bba WatchSource:0}: Error finding container 940ac92637132558cd794b90ed14e86c6d3c6fc9b5485452639b3a68c6c91bba: Status 404 returned error can't find the container with id 940ac92637132558cd794b90ed14e86c6d3c6fc9b5485452639b3a68c6c91bba Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.409779 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: W1124 13:46:50.410255 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0875a754_321a_4a3a_a798_941eefcdffe5.slice/crio-ab38c52f173f31c4ca0551690bd0ed42feeb5995274bf02549da4c14c1c01057 WatchSource:0}: Error finding container ab38c52f173f31c4ca0551690bd0ed42feeb5995274bf02549da4c14c1c01057: Status 404 returned error can't find the container with id ab38c52f173f31c4ca0551690bd0ed42feeb5995274bf02549da4c14c1c01057 Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.426829 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.443035 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.450374 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-gr7xg"] Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.451265 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-9jtl9"] Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.451639 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-msd5k"] Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.452905 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.453342 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.453664 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.455603 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.455935 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.456137 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.456339 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.456760 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.456988 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.457213 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.457405 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458639 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-mcd-auth-proxy-config\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458692 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458717 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r754h\" (UniqueName: \"kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458831 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-system-cni-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458879 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-binary-copy\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458908 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458932 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458958 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-proxy-tls\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.458981 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459004 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459022 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459026 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459065 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kmkb\" (UniqueName: \"kubernetes.io/projected/64b0c636-eda5-4094-b61e-26f6885f431b-kube-api-access-6kmkb\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459090 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459108 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459128 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459172 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459196 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459206 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r66kl\" (UniqueName: \"kubernetes.io/projected/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-kube-api-access-r66kl\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459265 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459293 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459313 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459337 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459374 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459425 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459449 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-cnibin\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459469 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459488 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459512 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-rootfs\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459531 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459562 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-os-release\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459599 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.459674 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.460189 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.460478 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.461130 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.461562 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.469486 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.469509 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.469618 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.469909 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.469970 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.482166 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.492120 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.505348 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.518805 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.529465 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.542050 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560138 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560173 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560190 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560206 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-mcd-auth-proxy-config\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560221 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-system-cni-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560237 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-binary-copy\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560251 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560264 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560279 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r754h\" (UniqueName: \"kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560301 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-proxy-tls\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560316 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560330 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560356 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kmkb\" (UniqueName: \"kubernetes.io/projected/64b0c636-eda5-4094-b61e-26f6885f431b-kube-api-access-6kmkb\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560371 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560386 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560400 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560417 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560432 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560449 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r66kl\" (UniqueName: \"kubernetes.io/projected/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-kube-api-access-r66kl\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560478 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560496 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560511 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560526 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560543 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560560 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560568 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560626 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-cnibin\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560591 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-cnibin\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560661 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560677 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560678 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560697 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560704 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-rootfs\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560721 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560745 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-os-release\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560842 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-os-release\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561494 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-tuning-conf-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561551 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561617 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64b0c636-eda5-4094-b61e-26f6885f431b-system-cni-dir\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561656 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561687 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.561759 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562292 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562395 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.560662 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562845 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562876 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562900 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.562922 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.563022 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.563069 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-rootfs\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.563093 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.563052 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.564050 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64b0c636-eda5-4094-b61e-26f6885f431b-cni-binary-copy\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.564553 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.564608 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.564644 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-mcd-auth-proxy-config\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.564868 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.565484 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.565662 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-proxy-tls\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.576810 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.577370 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kmkb\" (UniqueName: \"kubernetes.io/projected/64b0c636-eda5-4094-b61e-26f6885f431b-kube-api-access-6kmkb\") pod \"multus-additional-cni-plugins-gr7xg\" (UID: \"64b0c636-eda5-4094-b61e-26f6885f431b\") " pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.581768 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r754h\" (UniqueName: \"kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h\") pod \"ovnkube-node-msd5k\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.587926 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r66kl\" (UniqueName: \"kubernetes.io/projected/5dd7b91d-1aca-41aa-b8b4-ab97723e8074-kube-api-access-r66kl\") pod \"machine-config-daemon-9jtl9\" (UID: \"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\") " pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.590501 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.601174 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.618679 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bhjg8" event={"ID":"f2628d9f-c695-4d55-bf97-30c4443ca0d3","Type":"ContainerStarted","Data":"940ac92637132558cd794b90ed14e86c6d3c6fc9b5485452639b3a68c6c91bba"} Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.620258 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerStarted","Data":"4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641"} Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.620282 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerStarted","Data":"ab38c52f173f31c4ca0551690bd0ed42feeb5995274bf02549da4c14c1c01057"} Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.621727 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.635053 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.646555 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.657346 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.671228 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.683521 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.702987 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.713723 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.727795 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.750908 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.764695 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.776421 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.787398 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.789540 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.795433 4970 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.797182 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.797238 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.797253 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.797380 4970 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.801659 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" Nov 24 13:46:50 crc kubenswrapper[4970]: W1124 13:46:50.802657 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b38f92e_6c99_47b9_9079_0b0e455ddca1.slice/crio-9ca80d3c309203ce36121caca75393fc8f197217c0a4bf60d0cb60c84fb45dc3 WatchSource:0}: Error finding container 9ca80d3c309203ce36121caca75393fc8f197217c0a4bf60d0cb60c84fb45dc3: Status 404 returned error can't find the container with id 9ca80d3c309203ce36121caca75393fc8f197217c0a4bf60d0cb60c84fb45dc3 Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.806117 4970 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.806407 4970 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.807470 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.807739 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.807833 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.807909 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.808000 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.808097 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.813879 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:46:50 crc kubenswrapper[4970]: W1124 13:46:50.819620 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64b0c636_eda5_4094_b61e_26f6885f431b.slice/crio-160f3a0cfd69fb7f494c2ea77745185d03a8341c83ff9f693ce04b3710101276 WatchSource:0}: Error finding container 160f3a0cfd69fb7f494c2ea77745185d03a8341c83ff9f693ce04b3710101276: Status 404 returned error can't find the container with id 160f3a0cfd69fb7f494c2ea77745185d03a8341c83ff9f693ce04b3710101276 Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.822683 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.828733 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.833969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.834010 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.834019 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.834034 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.834043 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.835228 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.848358 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.851499 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.854800 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.854844 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.854856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.854872 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.854884 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.874759 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.879030 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.879073 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.879084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.879100 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.879111 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.897023 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.901965 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.902067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.902155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.902224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.902286 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.914321 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:50Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:50 crc kubenswrapper[4970]: E1124 13:46:50.914462 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.916797 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.916829 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.916837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.916851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:50 crc kubenswrapper[4970]: I1124 13:46:50.916861 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:50Z","lastTransitionTime":"2025-11-24T13:46:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.019027 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.019053 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.019061 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.019074 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.019092 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.121285 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.121538 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.121547 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.121561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.121570 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.223541 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.223592 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.223601 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.223615 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.223625 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.326016 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.326090 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.326103 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.326122 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.326133 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.428002 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.428051 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.428066 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.428090 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.428107 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.469445 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:51 crc kubenswrapper[4970]: E1124 13:46:51.469613 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.530541 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.530642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.530660 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.530686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.530704 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.625073 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bhjg8" event={"ID":"f2628d9f-c695-4d55-bf97-30c4443ca0d3","Type":"ContainerStarted","Data":"1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.627012 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.627046 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.627059 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"56487efbfcf767bff18249c3e5f84f8682b4a15c9d90ec1e7cc580b130772906"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.629281 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf" exitCode=0 Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.629347 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.629375 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerStarted","Data":"160f3a0cfd69fb7f494c2ea77745185d03a8341c83ff9f693ce04b3710101276"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633746 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" exitCode=0 Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633792 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633817 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"9ca80d3c309203ce36121caca75393fc8f197217c0a4bf60d0cb60c84fb45dc3"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633861 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633883 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633893 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.633926 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.643244 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.659904 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.681909 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.697196 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.710663 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.726189 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.737140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.737179 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.737187 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.737201 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.737211 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.740658 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.753325 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.766444 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.779085 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.792635 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.805471 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.816782 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.838699 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.840738 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.840763 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.840773 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.840787 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.840796 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.851260 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.900668 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.935039 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.942662 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.942697 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.942706 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.942720 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.942728 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:51Z","lastTransitionTime":"2025-11-24T13:46:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.950794 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.963679 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.982065 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:51 crc kubenswrapper[4970]: I1124 13:46:51.996962 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:51Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.009354 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.018875 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.031351 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.045659 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.045690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.045699 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.045712 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.045722 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.048352 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.059309 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.072599 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.093249 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.148501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.148540 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.148551 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.148566 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.148610 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.178015 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.178109 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.178138 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.178154 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.178184 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178276 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178329 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:00.178316685 +0000 UTC m=+35.466073978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178385 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:47:00.178377767 +0000 UTC m=+35.466135060 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178440 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178449 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178459 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178480 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:00.178474289 +0000 UTC m=+35.466231582 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178518 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178527 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178534 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178551 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:00.178546232 +0000 UTC m=+35.466303525 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178596 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.178614 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:00.178609073 +0000 UTC m=+35.466366366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.251217 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.251278 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.251292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.251310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.251324 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.354083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.354361 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.354424 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.354490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.354547 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.457363 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.457427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.457446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.457469 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.457486 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.469996 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.470106 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.470378 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:52 crc kubenswrapper[4970]: E1124 13:46:52.470436 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.559464 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.559770 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.559780 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.559794 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.559803 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.646166 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc" exitCode=0 Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.646470 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.658044 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.658078 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.658088 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.658096 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.658107 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.662451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.662489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.662500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.662515 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.662526 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.666971 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.682545 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.697076 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.717053 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.730291 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.741956 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.754750 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.766265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.766308 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.766321 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.766340 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.766352 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.770797 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.781283 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.792956 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.805006 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.816235 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.826915 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.843737 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.868405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.868441 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.868450 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.868462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.868471 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.971256 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.971310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.971327 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.971349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:52 crc kubenswrapper[4970]: I1124 13:46:52.971366 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:52Z","lastTransitionTime":"2025-11-24T13:46:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.073687 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.073737 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.073746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.073759 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.073768 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.176333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.176376 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.176387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.176407 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.176418 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.280289 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.280368 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.280397 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.280429 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.280451 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.383539 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.383617 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.383635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.383657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.383674 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.470030 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:53 crc kubenswrapper[4970]: E1124 13:46:53.470267 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.486375 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.486433 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.486450 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.486473 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.486492 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.589594 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.589639 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.589650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.589671 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.589690 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.663156 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163" exitCode=0 Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.663254 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.667700 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.685555 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.692840 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.692893 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.692910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.692933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.692950 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.703677 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.719057 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.746486 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.761074 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.772868 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.784840 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.795699 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.795734 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.795744 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.795758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.795768 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.801246 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.811420 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tc8pw"] Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.812113 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.812810 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.814069 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.814215 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.814363 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.815018 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.825396 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.838378 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.853277 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.864878 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.884383 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.895000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2bfd2947-8ebd-4cfa-b412-eec444657731-host\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.895047 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jpxx\" (UniqueName: \"kubernetes.io/projected/2bfd2947-8ebd-4cfa-b412-eec444657731-kube-api-access-6jpxx\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.895084 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2bfd2947-8ebd-4cfa-b412-eec444657731-serviceca\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.896340 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.897533 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.897559 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.897568 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.897600 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.897610 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.906671 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.919827 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.934486 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.947857 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.962743 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.981915 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.992149 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:53Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.996217 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2bfd2947-8ebd-4cfa-b412-eec444657731-host\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.996257 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jpxx\" (UniqueName: \"kubernetes.io/projected/2bfd2947-8ebd-4cfa-b412-eec444657731-kube-api-access-6jpxx\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.996289 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2bfd2947-8ebd-4cfa-b412-eec444657731-serviceca\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.996372 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2bfd2947-8ebd-4cfa-b412-eec444657731-host\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.997532 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2bfd2947-8ebd-4cfa-b412-eec444657731-serviceca\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.999522 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.999601 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.999625 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.999648 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:53 crc kubenswrapper[4970]: I1124 13:46:53.999667 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:53Z","lastTransitionTime":"2025-11-24T13:46:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.004549 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.017399 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.017704 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jpxx\" (UniqueName: \"kubernetes.io/projected/2bfd2947-8ebd-4cfa-b412-eec444657731-kube-api-access-6jpxx\") pod \"node-ca-tc8pw\" (UID: \"2bfd2947-8ebd-4cfa-b412-eec444657731\") " pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.034979 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.049168 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.060806 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.072995 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.093921 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.101265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.101448 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.101521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.101608 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.101693 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.143313 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tc8pw" Nov 24 13:46:54 crc kubenswrapper[4970]: W1124 13:46:54.160570 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bfd2947_8ebd_4cfa_b412_eec444657731.slice/crio-1c8e8bb25abca1ef47188a86265bcbaa160409e85c925ab73d75ed9bcaa5e0a6 WatchSource:0}: Error finding container 1c8e8bb25abca1ef47188a86265bcbaa160409e85c925ab73d75ed9bcaa5e0a6: Status 404 returned error can't find the container with id 1c8e8bb25abca1ef47188a86265bcbaa160409e85c925ab73d75ed9bcaa5e0a6 Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.204826 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.204881 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.204895 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.204913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.204926 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.307290 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.307525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.307534 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.307548 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.307557 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.410828 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.410882 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.410901 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.410924 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.410941 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.470070 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:54 crc kubenswrapper[4970]: E1124 13:46:54.470246 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.470287 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:54 crc kubenswrapper[4970]: E1124 13:46:54.470488 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.513206 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.513243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.513260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.513278 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.513292 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.616144 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.616227 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.616253 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.616286 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.616309 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.675130 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tc8pw" event={"ID":"2bfd2947-8ebd-4cfa-b412-eec444657731","Type":"ContainerStarted","Data":"28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.675236 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tc8pw" event={"ID":"2bfd2947-8ebd-4cfa-b412-eec444657731","Type":"ContainerStarted","Data":"1c8e8bb25abca1ef47188a86265bcbaa160409e85c925ab73d75ed9bcaa5e0a6"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.678562 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258" exitCode=0 Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.678659 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.700366 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.719031 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.719083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.719098 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.719118 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.719133 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.727425 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.744257 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.760472 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.780215 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.794464 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.807960 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822102 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822275 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822298 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.822340 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.834065 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.855917 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.874854 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.904513 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.924229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.924263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.924274 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.924299 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.924312 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:54Z","lastTransitionTime":"2025-11-24T13:46:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.925941 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.939209 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.950749 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.962535 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.975471 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:54 crc kubenswrapper[4970]: I1124 13:46:54.995167 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:54Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.012455 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.027334 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.027387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.027399 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.027418 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.027431 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.029501 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.043460 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.055143 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.067349 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.077023 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.088180 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.099309 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.110591 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.121899 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.129786 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.129816 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.129827 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.129840 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.129849 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.137688 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.147483 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.232348 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.232392 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.232405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.232422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.232435 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.335788 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.335890 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.335905 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.335925 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.335952 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.438852 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.438898 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.438912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.438929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.438940 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.470045 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:55 crc kubenswrapper[4970]: E1124 13:46:55.470385 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.496269 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.519819 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.534407 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.541253 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.541298 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.541314 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.541336 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.541351 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.556662 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.587735 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.607184 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.618815 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.636207 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.646937 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.646977 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.646985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.646998 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.647007 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.654948 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.670843 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.684923 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61" exitCode=0 Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.684990 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.689785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.691049 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.711810 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.729395 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.746136 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.749248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.749286 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.749295 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.749311 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.749322 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.765240 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.777938 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.801389 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.820304 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.833996 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.851034 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.851089 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.851106 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.851130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.851148 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.852336 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.869816 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.886083 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.901354 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.915902 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.930658 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.946754 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.953953 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.953989 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.954004 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.954024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.954038 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:55Z","lastTransitionTime":"2025-11-24T13:46:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.959769 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.974799 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:55 crc kubenswrapper[4970]: I1124 13:46:55.994751 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.004800 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.056322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.056360 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.056373 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.056389 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.056400 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.158757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.158814 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.158832 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.158854 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.158872 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.262239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.262293 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.262310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.262424 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.262446 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.365884 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.365939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.365956 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.365983 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.366001 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469376 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: E1124 13:46:56.469504 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469523 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469542 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469559 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469394 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.469600 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: E1124 13:46:56.469719 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.571893 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.571979 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.571997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.572019 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.572073 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.675526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.675623 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.675650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.675678 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.675699 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.697389 4970 generic.go:334] "Generic (PLEG): container finished" podID="64b0c636-eda5-4094-b61e-26f6885f431b" containerID="eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d" exitCode=0 Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.697462 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerDied","Data":"eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.723938 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.748569 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.764193 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.778810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.778847 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.778859 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.778877 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.778893 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.789822 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.810086 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.830842 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.843987 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.856159 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.867927 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881237 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881299 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881453 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881471 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.881484 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.895800 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.911208 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.928875 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.944031 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.963225 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:56Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.983221 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.983268 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.983281 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.983299 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:56 crc kubenswrapper[4970]: I1124 13:46:56.983311 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:56Z","lastTransitionTime":"2025-11-24T13:46:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.084953 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.085000 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.085015 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.085036 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.085047 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.189029 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.189097 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.189115 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.189141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.189159 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.292559 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.292688 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.292707 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.292732 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.292750 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.396229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.396327 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.396422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.396446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.396464 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.469464 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:57 crc kubenswrapper[4970]: E1124 13:46:57.469620 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.499566 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.499684 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.499700 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.499729 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.499745 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.603003 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.603049 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.603064 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.603083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.603097 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.705420 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.705465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.705521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.705545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.705563 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.706375 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" event={"ID":"64b0c636-eda5-4094-b61e-26f6885f431b","Type":"ContainerStarted","Data":"5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.712557 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.712976 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.713015 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.730072 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.750770 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.774451 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.783171 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.783334 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.797912 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.808511 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.808616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.808629 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.808651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.808670 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.814503 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.828099 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.846785 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.865033 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.877553 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.891346 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.903671 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.910793 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.910836 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.910848 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.910866 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.910880 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:57Z","lastTransitionTime":"2025-11-24T13:46:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.915853 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.930795 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.949221 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.960091 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.977627 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:57 crc kubenswrapper[4970]: I1124 13:46:57.990403 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:57Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.013848 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.013887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.013899 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.013917 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.013931 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.015836 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.027274 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.040118 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.052010 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.063207 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.074157 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.087370 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.105516 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.114593 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.116106 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.116149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.116160 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.116176 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.116189 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.127014 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.138786 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.153461 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.169840 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:46:58Z is after 2025-08-24T17:21:41Z" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.218702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.218760 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.218777 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.218799 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.218818 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.322344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.322401 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.322419 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.322448 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.322470 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.425385 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.425446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.425462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.425488 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.425506 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.469400 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.469417 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:46:58 crc kubenswrapper[4970]: E1124 13:46:58.469613 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:46:58 crc kubenswrapper[4970]: E1124 13:46:58.469798 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.529264 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.529326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.529344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.529372 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.529389 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.632830 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.632904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.632930 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.632960 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.633024 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.717360 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.736562 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.736652 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.736676 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.736705 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.736725 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.839931 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.839986 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.840003 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.840024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.840042 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.942130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.942166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.942175 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.942188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:58 crc kubenswrapper[4970]: I1124 13:46:58.942197 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:58Z","lastTransitionTime":"2025-11-24T13:46:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.045773 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.045842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.045860 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.045886 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.045913 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.149526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.149616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.149635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.149661 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.149678 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.253800 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.253854 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.253874 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.253898 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.253916 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.356501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.356562 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.356616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.356658 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.356679 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.459723 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.459762 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.459775 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.459790 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.459799 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.470373 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:46:59 crc kubenswrapper[4970]: E1124 13:46:59.470539 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.562278 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.562315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.562326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.562342 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.562354 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.664702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.664757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.664773 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.664792 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.664806 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.720735 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.767541 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.767598 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.767608 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.767624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.767637 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.870440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.870511 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.870533 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.870560 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.870613 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.988078 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.988125 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.988136 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.988150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:46:59 crc kubenswrapper[4970]: I1124 13:46:59.988159 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:46:59Z","lastTransitionTime":"2025-11-24T13:46:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.089904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.089933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.089941 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.089954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.089963 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.189300 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.189405 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.189440 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189476 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:47:16.189438684 +0000 UTC m=+51.477196017 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.189538 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189594 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189623 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189634 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189665 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189682 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.189688 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189639 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189759 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:16.189737611 +0000 UTC m=+51.477494984 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189794 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:16.189779262 +0000 UTC m=+51.477536645 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189802 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189734 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189849 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:16.189837144 +0000 UTC m=+51.477594447 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.189866 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:16.189858084 +0000 UTC m=+51.477615387 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.191380 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.191422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.191440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.191463 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.191479 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.294079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.294126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.294138 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.294155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.294167 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.397204 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.397258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.397276 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.397298 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.397316 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.469902 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.469982 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.470095 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:00 crc kubenswrapper[4970]: E1124 13:47:00.470233 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.499709 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.499765 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.499778 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.499800 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.499817 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.603033 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.603123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.603172 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.603199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.603222 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.701023 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.706617 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.706664 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.706681 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.706702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.706719 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.727514 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/0.log" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.731754 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465" exitCode=1 Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.731833 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.733209 4970 scope.go:117] "RemoveContainer" containerID="02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.754787 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.779998 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.799162 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.814236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.814292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.814309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.814333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.814349 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.819917 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.842832 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.857511 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.873397 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.904482 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:00Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:47:00.220210 6307 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:47:00.220233 6307 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:47:00.220270 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 13:47:00.220288 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:47:00.220305 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:47:00.220334 6307 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:47:00.220359 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:47:00.220371 6307 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 13:47:00.220386 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:47:00.220410 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:47:00.220425 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:47:00.220446 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:47:00.220457 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:47:00.220475 6307 factory.go:656] Stopping watch factory\\\\nI1124 13:47:00.220496 6307 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.917230 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.917300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.917319 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.917344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.917363 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:00Z","lastTransitionTime":"2025-11-24T13:47:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.924886 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.947601 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:00 crc kubenswrapper[4970]: I1124 13:47:00.968891 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:00Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.005351 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.021107 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.021151 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.021167 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.021187 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.021204 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.028735 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.051846 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.073559 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.124241 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.124304 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.124323 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.124349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.124367 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.126224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.126280 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.126300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.126324 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.126341 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.148664 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.154036 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.154101 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.154123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.154149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.154167 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.176877 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.181971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.182022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.182040 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.182063 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.182080 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.206010 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.211114 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.211195 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.211219 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.211251 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.211270 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.231541 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.236217 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.236262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.236279 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.236300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.236316 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.259332 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.259517 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.261966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.262001 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.262016 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.262039 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.262055 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.365067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.365109 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.365120 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.365138 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.365150 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.467107 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.467154 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.467163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.467179 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.467188 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.469414 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:01 crc kubenswrapper[4970]: E1124 13:47:01.469509 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.570185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.570225 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.570236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.570254 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.570265 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.672994 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.673032 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.673044 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.673060 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.673071 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.738209 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/0.log" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.741020 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.741398 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.763318 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.775148 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.775191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.775206 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.775226 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.775242 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.780816 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.792178 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.806244 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.818565 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.834488 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.851901 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.869239 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.877882 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.877928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.877939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.877959 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.877971 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.898740 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:00Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:47:00.220210 6307 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:47:00.220233 6307 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:47:00.220270 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 13:47:00.220288 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:47:00.220305 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:47:00.220334 6307 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:47:00.220359 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:47:00.220371 6307 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 13:47:00.220386 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:47:00.220410 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:47:00.220425 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:47:00.220446 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:47:00.220457 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:47:00.220475 6307 factory.go:656] Stopping watch factory\\\\nI1124 13:47:00.220496 6307 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.924367 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.941318 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.957317 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.973868 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.979796 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.979824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.979834 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.979851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.979865 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:01Z","lastTransitionTime":"2025-11-24T13:47:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:01 crc kubenswrapper[4970]: I1124 13:47:01.987834 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.001343 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:01Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.026141 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76"] Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.026724 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.028142 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.029848 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.042723 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.055385 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.076745 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.083118 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.083149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.083163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.083177 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.083186 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.097532 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.106885 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h98nv\" (UniqueName: \"kubernetes.io/projected/642b8bf5-104e-4448-9f11-8c2dd5d12c41-kube-api-access-h98nv\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.106939 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.107033 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.107072 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.112485 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.126079 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.140151 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.154806 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.167980 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.177520 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.185776 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.185816 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.185828 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.185846 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.185860 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.189825 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.206167 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.208048 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.208292 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.208632 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h98nv\" (UniqueName: \"kubernetes.io/projected/642b8bf5-104e-4448-9f11-8c2dd5d12c41-kube-api-access-h98nv\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.208694 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.209190 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.209280 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.215694 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/642b8bf5-104e-4448-9f11-8c2dd5d12c41-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.226190 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.237509 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h98nv\" (UniqueName: \"kubernetes.io/projected/642b8bf5-104e-4448-9f11-8c2dd5d12c41-kube-api-access-h98nv\") pod \"ovnkube-control-plane-749d76644c-4nq76\" (UID: \"642b8bf5-104e-4448-9f11-8c2dd5d12c41\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.243846 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.276460 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:00Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:47:00.220210 6307 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:47:00.220233 6307 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:47:00.220270 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 13:47:00.220288 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:47:00.220305 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:47:00.220334 6307 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:47:00.220359 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:47:00.220371 6307 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 13:47:00.220386 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:47:00.220410 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:47:00.220425 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:47:00.220446 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:47:00.220457 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:47:00.220475 6307 factory.go:656] Stopping watch factory\\\\nI1124 13:47:00.220496 6307 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.289070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.289130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.289153 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.289184 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.289207 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.291990 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.341659 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" Nov 24 13:47:02 crc kubenswrapper[4970]: W1124 13:47:02.361017 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod642b8bf5_104e_4448_9f11_8c2dd5d12c41.slice/crio-d8cea5adfa8389a917514cf5d86ae2e119234d9965f0f5f6321d90cd3b98a025 WatchSource:0}: Error finding container d8cea5adfa8389a917514cf5d86ae2e119234d9965f0f5f6321d90cd3b98a025: Status 404 returned error can't find the container with id d8cea5adfa8389a917514cf5d86ae2e119234d9965f0f5f6321d90cd3b98a025 Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.392455 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.392536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.392560 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.392626 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.392653 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.470091 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:02 crc kubenswrapper[4970]: E1124 13:47:02.470294 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.470877 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:02 crc kubenswrapper[4970]: E1124 13:47:02.471287 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.496070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.496136 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.496155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.496180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.496197 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.598853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.598914 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.598932 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.598955 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.598971 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.701567 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.701657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.701674 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.701697 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.701749 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.746123 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/1.log" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.747338 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/0.log" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.751957 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa" exitCode=1 Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.752048 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.752112 4970 scope.go:117] "RemoveContainer" containerID="02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.753287 4970 scope.go:117] "RemoveContainer" containerID="e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa" Nov 24 13:47:02 crc kubenswrapper[4970]: E1124 13:47:02.753650 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.756447 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" event={"ID":"642b8bf5-104e-4448-9f11-8c2dd5d12c41","Type":"ContainerStarted","Data":"810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.756512 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" event={"ID":"642b8bf5-104e-4448-9f11-8c2dd5d12c41","Type":"ContainerStarted","Data":"d8cea5adfa8389a917514cf5d86ae2e119234d9965f0f5f6321d90cd3b98a025"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.779041 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.797720 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.804650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.804701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.804719 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.804741 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.804758 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.814931 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.844352 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.863888 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.877670 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.894495 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.907556 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.907614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.907624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.907641 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.907654 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:02Z","lastTransitionTime":"2025-11-24T13:47:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.918536 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.931877 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.948899 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.969012 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02ff59fd2344cccd0171b7db36b5f73200c3a28986b611c7b5953dc5e8613465\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:00Z\\\",\\\"message\\\":\\\"/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 13:47:00.220210 6307 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 13:47:00.220233 6307 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 13:47:00.220270 6307 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 13:47:00.220288 6307 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 13:47:00.220305 6307 handler.go:208] Removed *v1.Node event handler 7\\\\nI1124 13:47:00.220334 6307 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1124 13:47:00.220359 6307 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 13:47:00.220371 6307 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 13:47:00.220386 6307 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1124 13:47:00.220410 6307 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 13:47:00.220425 6307 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 13:47:00.220446 6307 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 13:47:00.220457 6307 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 13:47:00.220475 6307 factory.go:656] Stopping watch factory\\\\nI1124 13:47:00.220496 6307 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.980428 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:02 crc kubenswrapper[4970]: I1124 13:47:02.993042 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.003516 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.010216 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.010269 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.010285 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.010307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.010323 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.016399 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.027977 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.119097 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.119150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.119162 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.119180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.119192 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.221755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.221822 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.221847 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.221878 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.221898 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.324872 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.324917 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.324933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.324954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.324970 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.428135 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.428198 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.428220 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.428245 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.428263 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.470124 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:03 crc kubenswrapper[4970]: E1124 13:47:03.470327 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.531490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.531558 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.531613 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.531644 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.531664 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.635509 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.635627 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.635655 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.635699 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.635719 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.738709 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.738779 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.738796 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.738820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.738839 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.765528 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/1.log" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.771530 4970 scope.go:117] "RemoveContainer" containerID="e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa" Nov 24 13:47:03 crc kubenswrapper[4970]: E1124 13:47:03.771727 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.772917 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" event={"ID":"642b8bf5-104e-4448-9f11-8c2dd5d12c41","Type":"ContainerStarted","Data":"b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.784314 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.799970 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.813556 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.830870 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.841028 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.841082 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.841124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.841150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.841167 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.848211 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.869439 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.888554 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.905314 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.924238 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.936546 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kv7zv"] Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.937342 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:03 crc kubenswrapper[4970]: E1124 13:47:03.937448 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.944897 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.944940 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.944958 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.944979 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.944998 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:03Z","lastTransitionTime":"2025-11-24T13:47:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.955552 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:03 crc kubenswrapper[4970]: I1124 13:47:03.981939 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.007133 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.023272 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.026117 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjt8w\" (UniqueName: \"kubernetes.io/projected/8b0fceff-420b-4465-9afa-99af03170306-kube-api-access-kjt8w\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.026245 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.040716 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.047900 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.047939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.047954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.047975 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.047990 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.060261 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.079170 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.102247 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.127553 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjt8w\" (UniqueName: \"kubernetes.io/projected/8b0fceff-420b-4465-9afa-99af03170306-kube-api-access-kjt8w\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.127760 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.128021 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.128147 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:04.628113326 +0000 UTC m=+39.915870699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.132422 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151439 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151522 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151554 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151613 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.151727 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.161842 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjt8w\" (UniqueName: \"kubernetes.io/projected/8b0fceff-420b-4465-9afa-99af03170306-kube-api-access-kjt8w\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.172717 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.195440 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.217341 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.236861 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.255130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.255194 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.255208 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.255229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.255244 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.267869 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.282170 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.297528 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.316384 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.339446 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.357725 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.357767 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.357778 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.357795 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.357807 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.364737 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.387760 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.409731 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.430943 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.448761 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:04Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.460814 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.460851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.460862 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.460876 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.460887 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.469685 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.469733 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.469820 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.469864 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.562982 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.563042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.563058 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.563082 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.563099 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.631775 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.631978 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:04 crc kubenswrapper[4970]: E1124 13:47:04.632087 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:05.6320649 +0000 UTC m=+40.919822203 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.666268 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.666323 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.666334 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.666352 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.666364 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.769713 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.769784 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.769802 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.769826 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.769843 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.872982 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.873030 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.873047 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.873070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.873086 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.976624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.976759 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.976819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.976899 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:04 crc kubenswrapper[4970]: I1124 13:47:04.976924 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:04Z","lastTransitionTime":"2025-11-24T13:47:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.080134 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.080308 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.080337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.080366 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.080388 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.183041 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.183105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.183123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.183149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.183166 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.286653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.286709 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.286729 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.286751 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.286768 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.390149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.390194 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.390213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.390239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.390255 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.469739 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:05 crc kubenswrapper[4970]: E1124 13:47:05.470070 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.470566 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:05 crc kubenswrapper[4970]: E1124 13:47:05.470744 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.492611 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.492653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.492665 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.492679 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.492690 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.501916 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.524708 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.546318 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.571848 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.592157 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.596317 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.596451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.596480 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.596528 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.596558 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.613176 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.635344 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.641296 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:05 crc kubenswrapper[4970]: E1124 13:47:05.641475 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:05 crc kubenswrapper[4970]: E1124 13:47:05.641564 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:07.641538646 +0000 UTC m=+42.929295969 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.650332 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.671996 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.688543 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.699309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.699350 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.699365 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.699385 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.699401 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.701880 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.716030 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.738122 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.751238 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.764166 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.783963 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.804981 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.805289 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.805616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.805905 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.805961 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.805986 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.909154 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.909197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.909209 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.909225 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:05 crc kubenswrapper[4970]: I1124 13:47:05.909237 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:05Z","lastTransitionTime":"2025-11-24T13:47:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.011690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.011737 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.011746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.011759 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.011767 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.114712 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.114857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.114884 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.114913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.114935 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.218024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.218070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.218081 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.218096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.218108 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.321269 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.321330 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.321344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.321378 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.321390 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.424658 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.424712 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.424731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.424758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.424775 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.470024 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.470052 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:06 crc kubenswrapper[4970]: E1124 13:47:06.470223 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:06 crc kubenswrapper[4970]: E1124 13:47:06.470381 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.527572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.527651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.527663 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.527680 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.527690 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.630810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.630887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.630912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.630942 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.630965 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.734159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.734222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.734245 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.734277 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.734300 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.837528 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.837631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.837655 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.837679 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.837696 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.941439 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.941526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.941561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.941622 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:06 crc kubenswrapper[4970]: I1124 13:47:06.941644 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:06Z","lastTransitionTime":"2025-11-24T13:47:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.044086 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.044147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.044166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.044213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.044251 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.147284 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.147336 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.147352 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.147378 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.147397 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.249849 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.249887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.249896 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.249910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.249919 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.352443 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.352513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.352549 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.352624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.352672 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.456272 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.456346 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.456364 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.456388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.456405 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.470146 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:07 crc kubenswrapper[4970]: E1124 13:47:07.470296 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.470147 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:07 crc kubenswrapper[4970]: E1124 13:47:07.470481 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.559155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.559222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.559243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.559269 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.559286 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662134 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662449 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662474 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662482 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662496 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: E1124 13:47:07.662495 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.662505 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: E1124 13:47:07.662631 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:11.662601401 +0000 UTC m=+46.950358734 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.765300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.765377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.765399 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.765427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.765448 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.868390 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.868426 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.868435 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.868449 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.868458 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.970401 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.970461 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.970494 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.970521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:07 crc kubenswrapper[4970]: I1124 13:47:07.970541 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:07Z","lastTransitionTime":"2025-11-24T13:47:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.073535 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.073660 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.073740 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.073770 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.073787 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.176757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.176807 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.176820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.176838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.176855 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.280624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.280728 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.280748 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.280772 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.280829 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.383881 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.383934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.383951 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.383972 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.383990 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.469544 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.469662 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:08 crc kubenswrapper[4970]: E1124 13:47:08.469750 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:08 crc kubenswrapper[4970]: E1124 13:47:08.469830 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.487113 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.487349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.487374 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.487401 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.487419 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.590864 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.590913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.590929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.590950 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.590966 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.694220 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.694277 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.694294 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.694318 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.694335 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.796470 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.796550 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.796606 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.796640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.796664 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.899987 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.900036 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.900047 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.900066 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:08 crc kubenswrapper[4970]: I1124 13:47:08.900077 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:08Z","lastTransitionTime":"2025-11-24T13:47:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.004139 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.004243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.004262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.004288 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.004304 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.107870 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.107930 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.107947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.107970 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.107988 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.210544 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.210600 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.210612 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.210626 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.210634 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.313129 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.313217 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.313240 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.313265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.313285 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.416783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.416875 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.416906 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.416937 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.416959 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.469367 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:09 crc kubenswrapper[4970]: E1124 13:47:09.469536 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.469382 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:09 crc kubenswrapper[4970]: E1124 13:47:09.469804 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.519436 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.519490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.519505 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.519525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.519542 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.622823 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.622887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.622905 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.622929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.622946 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.726775 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.726865 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.726889 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.726918 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.726937 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.830344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.830394 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.830405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.830423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.830440 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.933453 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.933530 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.933552 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.933616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:09 crc kubenswrapper[4970]: I1124 13:47:09.933668 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:09Z","lastTransitionTime":"2025-11-24T13:47:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.042695 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.042747 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.042761 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.042783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.042799 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.145972 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.146046 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.146072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.146102 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.146125 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.250689 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.250749 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.250768 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.250791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.250808 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.353931 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.354002 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.354024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.354049 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.354066 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.457799 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.457881 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.457910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.457940 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.457960 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.470345 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:10 crc kubenswrapper[4970]: E1124 13:47:10.470492 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.470348 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:10 crc kubenswrapper[4970]: E1124 13:47:10.470727 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.561008 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.561079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.561101 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.561130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.561153 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.665634 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.665682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.665698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.665721 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.665738 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.768885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.768929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.768943 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.768959 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.768971 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.874131 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.874205 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.874222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.874247 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.874263 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.977868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.978026 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.978044 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.978069 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:10 crc kubenswrapper[4970]: I1124 13:47:10.978086 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:10Z","lastTransitionTime":"2025-11-24T13:47:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.081563 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.081631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.081640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.081657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.081666 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.184202 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.184237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.184259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.184272 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.184281 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.287077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.287153 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.287174 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.287199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.287216 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.355616 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.355698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.355733 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.355768 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.355790 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.377122 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.382541 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.382640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.382657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.382682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.382699 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.403035 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.408716 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.408769 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.408789 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.408813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.408830 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.429665 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.434654 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.434752 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.434772 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.434796 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.434813 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.454393 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.459212 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.459259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.459279 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.459310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.459333 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.470344 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.470509 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.471018 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.471175 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.480605 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:11Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.480802 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.482442 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.482491 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.482507 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.482530 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.482620 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.585160 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.585226 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.585245 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.585275 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.585311 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.688688 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.688757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.688774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.688798 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.688815 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.708841 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.709788 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:11 crc kubenswrapper[4970]: E1124 13:47:11.709920 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:19.709868463 +0000 UTC m=+54.997625796 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.791508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.791569 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.791639 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.791671 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.791692 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.895514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.895609 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.895627 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.895651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.895671 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.999133 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.999196 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.999219 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.999246 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:11 crc kubenswrapper[4970]: I1124 13:47:11.999265 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:11Z","lastTransitionTime":"2025-11-24T13:47:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.101935 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.101999 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.102021 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.102049 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.102075 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.205892 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.205946 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.205964 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.205998 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.206035 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.308990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.309042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.309059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.309087 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.309108 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.412139 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.412188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.412205 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.412229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.412247 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.470421 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.470426 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:12 crc kubenswrapper[4970]: E1124 13:47:12.470653 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:12 crc kubenswrapper[4970]: E1124 13:47:12.470801 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.514832 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.514908 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.514929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.514957 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.514974 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.618402 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.618458 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.618475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.618501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.618519 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.721545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.721641 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.721666 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.721696 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.721721 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.824245 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.824305 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.824322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.824345 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.824366 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.927846 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.927915 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.927937 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.927971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:12 crc kubenswrapper[4970]: I1124 13:47:12.927994 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:12Z","lastTransitionTime":"2025-11-24T13:47:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.031619 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.031694 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.031715 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.031743 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.031763 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.135421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.135475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.135491 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.135514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.135533 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.239129 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.239214 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.239233 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.239260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.239279 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.342256 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.342333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.342356 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.342388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.342410 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.444908 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.444989 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.445006 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.445031 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.445051 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.470285 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.470415 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:13 crc kubenswrapper[4970]: E1124 13:47:13.470621 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:13 crc kubenswrapper[4970]: E1124 13:47:13.470727 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.548365 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.548423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.548443 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.548467 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.548483 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.651417 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.651475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.651512 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.651543 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.651565 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.754853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.755196 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.755342 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.755508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.755746 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.859482 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.859544 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.859562 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.859631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.859667 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.962092 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.962136 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.962146 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.962159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:13 crc kubenswrapper[4970]: I1124 13:47:13.962170 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:13Z","lastTransitionTime":"2025-11-24T13:47:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.064919 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.065232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.065325 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.065423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.065509 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.169873 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.170249 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.170436 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.170715 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.170928 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.274888 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.274971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.274997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.275032 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.275056 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.378460 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.378876 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.379194 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.379783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.380005 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.470062 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:14 crc kubenswrapper[4970]: E1124 13:47:14.470443 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.470816 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:14 crc kubenswrapper[4970]: E1124 13:47:14.470926 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.485156 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.485246 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.485271 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.485306 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.485331 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.588360 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.588449 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.588471 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.588498 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.588517 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.691636 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.691672 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.691681 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.691696 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.691706 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.794041 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.794088 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.794104 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.794126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.794139 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.897199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.897286 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.897316 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.897349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:14 crc kubenswrapper[4970]: I1124 13:47:14.897369 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:14Z","lastTransitionTime":"2025-11-24T13:47:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.000818 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.000899 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.000922 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.000947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.000965 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.103850 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.103912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.103933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.103955 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.103969 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.206966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.207054 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.207070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.207091 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.207104 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.310666 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.310740 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.310752 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.310775 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.310789 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.414125 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.414179 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.414189 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.414207 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.414218 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.470458 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:15 crc kubenswrapper[4970]: E1124 13:47:15.470723 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.470764 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:15 crc kubenswrapper[4970]: E1124 13:47:15.471082 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.484337 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.498682 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.513992 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.517337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.517388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.517405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.517434 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.517450 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.527746 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.542664 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.564357 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.577817 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.588945 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.604021 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620157 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620365 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620398 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620412 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620436 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.620455 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.639597 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.656401 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.670089 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.685729 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.697666 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.720132 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.728240 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.728299 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.728313 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.728336 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.728353 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.734977 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:15Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.832555 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.832628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.832641 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.832659 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.832671 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.935266 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.935321 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.935339 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.935363 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:15 crc kubenswrapper[4970]: I1124 13:47:15.935380 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:15Z","lastTransitionTime":"2025-11-24T13:47:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.038850 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.038910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.038927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.038950 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.038968 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.141879 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.141935 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.141952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.141975 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.141993 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.246201 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.246313 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.246341 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.246381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.246411 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.257853 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.258070 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258174 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:47:48.258077072 +0000 UTC m=+83.545834465 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.258282 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258331 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258378 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258406 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.258501 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258522 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:48.258496423 +0000 UTC m=+83.546253756 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.258645 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258887 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258914 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258948 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258976 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.258974 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.259011 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:48.258982615 +0000 UTC m=+83.546740078 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.259061 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:48.259042336 +0000 UTC m=+83.546799659 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.259153 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:48.259135539 +0000 UTC m=+83.546892862 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.350969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.351101 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.351122 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.351150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.351168 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.454512 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.454673 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.454703 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.454743 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.454771 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.469890 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.469927 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.470071 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:16 crc kubenswrapper[4970]: E1124 13:47:16.470372 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.558145 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.558186 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.558199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.558217 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.558231 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.660957 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.661043 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.661059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.661086 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.661102 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.764918 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.764980 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.764990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.765010 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.765023 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.868401 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.868460 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.868475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.868498 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.868509 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.970828 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.970900 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.970911 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.970933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:16 crc kubenswrapper[4970]: I1124 13:47:16.970949 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:16Z","lastTransitionTime":"2025-11-24T13:47:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.074208 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.074271 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.074283 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.074298 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.074310 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.177508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.177651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.177686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.177716 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.177737 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.280944 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.281000 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.281018 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.281041 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.281058 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.384316 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.384392 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.384404 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.384419 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.384429 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.469618 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.469731 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:17 crc kubenswrapper[4970]: E1124 13:47:17.470358 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:17 crc kubenswrapper[4970]: E1124 13:47:17.470788 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.481140 4970 scope.go:117] "RemoveContainer" containerID="e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.487233 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.487326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.487346 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.487635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.487747 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.596820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.596886 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.596904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.596927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.596944 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.700174 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.700210 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.700223 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.700242 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.700259 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.802807 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.802854 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.802870 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.802894 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.802911 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.826487 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/1.log" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.828979 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.830253 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.846537 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.886257 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.905413 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.905474 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.905491 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.905514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.905530 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:17Z","lastTransitionTime":"2025-11-24T13:47:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.911049 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.932066 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.954729 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.978286 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:17 crc kubenswrapper[4970]: I1124 13:47:17.994562 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:17Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.009387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.009440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.009450 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.009468 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.009801 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.011962 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.024892 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.036437 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.049076 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.061621 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.074930 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.101139 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.112332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.112383 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.112396 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.112415 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.112427 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.113829 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.127047 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.142220 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.214139 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.214180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.214195 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.214213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.214224 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.316940 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.316985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.316996 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.317017 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.317027 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.419258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.419309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.419323 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.419340 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.419352 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.470294 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:18 crc kubenswrapper[4970]: E1124 13:47:18.470463 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.470294 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:18 crc kubenswrapper[4970]: E1124 13:47:18.470702 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.521682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.521766 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.521785 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.521807 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.521823 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.624536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.624625 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.624638 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.624655 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.624666 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.728250 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.728307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.728325 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.728349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.728367 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.832322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.832389 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.832405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.832437 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.832473 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.836900 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/2.log" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.837820 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/1.log" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.842814 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" exitCode=1 Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.842904 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.842983 4970 scope.go:117] "RemoveContainer" containerID="e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.843873 4970 scope.go:117] "RemoveContainer" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" Nov 24 13:47:18 crc kubenswrapper[4970]: E1124 13:47:18.844107 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.866234 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.878773 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.903249 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.921328 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.934712 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.934758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.934772 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.934793 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.934808 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:18Z","lastTransitionTime":"2025-11-24T13:47:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.939367 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.951816 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.964756 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.975777 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:18 crc kubenswrapper[4970]: I1124 13:47:18.991665 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:18Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.002719 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.016991 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.030288 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.037347 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.037399 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.037416 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.037441 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.037460 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.039950 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.056976 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.082085 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.097367 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.114486 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.140016 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.140047 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.140056 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.140068 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.140077 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.182409 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.195010 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.199823 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.213787 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.227682 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.239228 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.242169 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.242202 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.242214 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.242228 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.242237 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.253430 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.267351 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.284754 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.301843 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.316100 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.342737 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3caa30a25845fc67bc5a84019796048a6e9e7bf38b7272dd5ea6ec1fb34a7fa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"message\\\":\\\"ewall event handler 9\\\\nI1124 13:47:01.852392 6443 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1124 13:47:01.852452 6443 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852597 6443 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852641 6443 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 13:47:01.852705 6443 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852730 6443 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.852776 6443 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 13:47:01.853273 6443 ovnkube.go:599] Stopped ovnkube\\\\nI1124 13:47:01.853324 6443 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1124 13:47:01.853410 6443 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.345965 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.346034 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.346052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.346079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.346097 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.361132 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.374731 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.385431 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.400704 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.430875 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.449423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.449504 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.449528 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.449558 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.449604 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.452973 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.469913 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.469909 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:19 crc kubenswrapper[4970]: E1124 13:47:19.470079 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:19 crc kubenswrapper[4970]: E1124 13:47:19.470221 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.476267 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.552812 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.552894 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.552918 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.552949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.552978 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.656447 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.656534 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.656556 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.656630 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.656650 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.760302 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.760536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.760644 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.760680 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.760707 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.803686 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:19 crc kubenswrapper[4970]: E1124 13:47:19.803851 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:19 crc kubenswrapper[4970]: E1124 13:47:19.803929 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:47:35.803908783 +0000 UTC m=+71.091666086 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.849870 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/2.log" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.855824 4970 scope.go:117] "RemoveContainer" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" Nov 24 13:47:19 crc kubenswrapper[4970]: E1124 13:47:19.856144 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.863560 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.863665 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.863710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.863735 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.863755 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.874345 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.912014 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.933505 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.948128 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.964538 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.967236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.967270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.967282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.967298 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.967310 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:19Z","lastTransitionTime":"2025-11-24T13:47:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.979396 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:19 crc kubenswrapper[4970]: I1124 13:47:19.995570 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:19Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.015196 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.030645 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.050366 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070196 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070212 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070221 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.070711 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.089765 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.110364 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.130691 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.159470 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.172868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.172921 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.172939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.172962 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.172979 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.185750 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.200127 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.222126 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:20Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.275063 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.275145 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.275168 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.275198 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.275221 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.378248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.378318 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.378334 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.378356 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.378373 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.469663 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.469732 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:20 crc kubenswrapper[4970]: E1124 13:47:20.469855 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:20 crc kubenswrapper[4970]: E1124 13:47:20.470015 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.481761 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.481867 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.481921 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.481946 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.481969 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.584405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.584471 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.584489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.584513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.584534 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.687603 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.687673 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.687685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.687705 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.687718 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.790712 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.790824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.790838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.790862 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.790880 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.894332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.894414 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.894438 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.894468 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.894493 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.997448 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.997526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.997553 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.997614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:20 crc kubenswrapper[4970]: I1124 13:47:20.997643 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:20Z","lastTransitionTime":"2025-11-24T13:47:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.101248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.101307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.101326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.101350 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.101367 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.204067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.204124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.204140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.204163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.204179 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.307490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.307568 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.307622 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.307648 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.307666 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.411098 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.411165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.411188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.411219 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.411242 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.470156 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.470171 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.470351 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.470518 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.514095 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.514177 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.514203 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.514234 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.514258 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.617455 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.617519 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.617536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.617561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.617621 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.720039 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.720145 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.720628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.720690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.720710 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.823851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.823909 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.823925 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.823947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.823964 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.834508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.834679 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.834707 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.834738 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.834767 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.858444 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.866675 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.866775 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.866802 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.866824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.866847 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.884083 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.888511 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.888556 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.888568 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.888615 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.888628 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.902535 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.908090 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.908128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.908136 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.908150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.908159 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.919476 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.924100 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.924157 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.924170 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.924185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.924195 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.940513 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:21Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:21 crc kubenswrapper[4970]: E1124 13:47:21.940640 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.941888 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.941912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.941929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.941943 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:21 crc kubenswrapper[4970]: I1124 13:47:21.941954 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:21Z","lastTransitionTime":"2025-11-24T13:47:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.044147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.044197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.044210 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.044227 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.044242 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.147256 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.147340 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.147365 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.147397 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.147420 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.250877 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.250936 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.250956 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.250983 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.251004 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.353767 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.353837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.353863 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.353892 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.353915 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.456960 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.457035 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.457056 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.457080 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.457097 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.470424 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.470444 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:22 crc kubenswrapper[4970]: E1124 13:47:22.470648 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:22 crc kubenswrapper[4970]: E1124 13:47:22.470763 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.559976 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.560052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.560075 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.560105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.560130 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.663457 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.663546 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.663663 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.663705 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.663727 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.766270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.766318 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.766331 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.766346 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.766357 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.869011 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.869072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.869092 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.869118 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.869140 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.972092 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.972483 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.972500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.972525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:22 crc kubenswrapper[4970]: I1124 13:47:22.972542 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:22Z","lastTransitionTime":"2025-11-24T13:47:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.075451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.075529 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.075547 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.075572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.075808 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.179112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.179184 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.179210 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.179239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.179260 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.282518 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.282609 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.282626 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.282648 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.282665 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.386113 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.386144 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.386154 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.386176 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.386187 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.469668 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.469765 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:23 crc kubenswrapper[4970]: E1124 13:47:23.469914 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:23 crc kubenswrapper[4970]: E1124 13:47:23.470045 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.489353 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.489641 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.489777 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.489919 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.490041 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.593087 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.593537 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.593951 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.594347 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.594726 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.697903 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.698347 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.698695 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.698950 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.699335 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.802663 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.802738 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.802755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.802778 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.802795 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.905803 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.905890 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.905913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.905944 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:23 crc kubenswrapper[4970]: I1124 13:47:23.905967 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:23Z","lastTransitionTime":"2025-11-24T13:47:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.008889 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.008952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.008969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.008992 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.009011 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.112803 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.112874 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.112899 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.112927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.112949 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.215552 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.215642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.215660 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.215684 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.215703 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.318440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.318490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.318501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.318522 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.318537 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.421241 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.421347 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.421374 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.421400 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.421417 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.470166 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.470266 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:24 crc kubenswrapper[4970]: E1124 13:47:24.470357 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:24 crc kubenswrapper[4970]: E1124 13:47:24.470410 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.524949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.525008 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.525040 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.525077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.525124 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.628051 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.628133 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.628154 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.628180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.628202 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.731749 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.731798 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.731810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.731827 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.731843 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.835495 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.835570 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.835635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.835667 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.835690 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.938508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.938606 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.938624 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.938643 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:24 crc kubenswrapper[4970]: I1124 13:47:24.938659 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:24Z","lastTransitionTime":"2025-11-24T13:47:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.041156 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.041237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.041259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.041294 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.041316 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.145038 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.145232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.145266 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.145349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.145417 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.250156 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.250364 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.250394 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.250423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.250444 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.353653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.353720 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.353746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.353776 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.353799 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.457109 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.457178 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.457195 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.457219 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.457240 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.469434 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.469517 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:25 crc kubenswrapper[4970]: E1124 13:47:25.469621 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:25 crc kubenswrapper[4970]: E1124 13:47:25.469759 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.497167 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.514812 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.531461 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.548449 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.559384 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.559450 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.559468 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.559484 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.559496 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.565814 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.601022 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.618343 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.632077 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.648483 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.662677 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.662796 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.662838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.662857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.662870 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.669820 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.688459 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.709883 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.741819 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.760051 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.765500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.765537 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.765553 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.765610 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.765628 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.777657 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.797643 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.818819 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.844253 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:25Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.868597 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.868631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.868640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.868654 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.868662 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.971485 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.971521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.971529 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.971544 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:25 crc kubenswrapper[4970]: I1124 13:47:25.971554 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:25Z","lastTransitionTime":"2025-11-24T13:47:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.074654 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.074751 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.074773 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.074802 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.074819 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.182748 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.182836 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.182856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.182882 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.182899 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.286289 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.286358 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.286376 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.286402 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.286425 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.391804 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.391895 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.391912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.391975 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.391995 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.470014 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:26 crc kubenswrapper[4970]: E1124 13:47:26.470251 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.470389 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:26 crc kubenswrapper[4970]: E1124 13:47:26.470752 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.494858 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.494903 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.494912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.494928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.494939 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.597345 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.597400 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.597413 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.597429 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.597466 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.700270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.700672 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.700836 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.700985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.701134 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.804018 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.804683 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.805069 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.805260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.805407 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.907791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.907831 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.907874 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.907892 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:26 crc kubenswrapper[4970]: I1124 13:47:26.907906 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:26Z","lastTransitionTime":"2025-11-24T13:47:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.010006 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.010080 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.010113 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.010143 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.010167 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.112837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.112871 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.112879 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.112890 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.112900 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.214956 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.215657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.215920 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.216126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.216239 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.319307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.319371 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.319398 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.319427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.319446 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.421676 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.421731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.421739 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.421756 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.421766 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.469671 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.469757 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:27 crc kubenswrapper[4970]: E1124 13:47:27.469861 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:27 crc kubenswrapper[4970]: E1124 13:47:27.469921 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.526051 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.526094 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.526105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.526121 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.526133 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.628848 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.628896 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.628908 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.628931 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.628953 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.731961 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.732007 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.732022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.732038 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.732051 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.834332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.834381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.834398 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.834419 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.834436 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.937044 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.937116 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.937128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.937143 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:27 crc kubenswrapper[4970]: I1124 13:47:27.937154 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:27Z","lastTransitionTime":"2025-11-24T13:47:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.040319 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.040362 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.040371 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.040385 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.040394 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.143278 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.143321 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.143334 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.143349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.143363 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.265434 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.265481 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.265490 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.265509 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.265518 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.367438 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.367509 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.367527 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.367553 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.367569 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.469374 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.469357 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:28 crc kubenswrapper[4970]: E1124 13:47:28.469657 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:28 crc kubenswrapper[4970]: E1124 13:47:28.469707 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.471127 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.471172 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.471181 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.471196 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.471207 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.574698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.574824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.574850 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.574876 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.575073 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.678972 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.679080 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.679141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.679370 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.679390 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.783473 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.783537 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.783556 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.783795 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.783823 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.885821 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.885857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.885866 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.885879 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.885888 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.988730 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.988774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.988785 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.988802 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:28 crc kubenswrapper[4970]: I1124 13:47:28.988813 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:28Z","lastTransitionTime":"2025-11-24T13:47:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.091706 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.091758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.091769 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.091786 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.091801 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.194673 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.194724 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.194748 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.194811 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.194835 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.298168 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.298221 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.298237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.298259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.298276 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.401013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.401086 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.401108 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.401134 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.401157 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.470281 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:29 crc kubenswrapper[4970]: E1124 13:47:29.470410 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.470293 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:29 crc kubenswrapper[4970]: E1124 13:47:29.471019 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.503745 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.503794 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.503803 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.503819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.503828 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.606375 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.606417 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.606426 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.606440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.606451 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.708254 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.708305 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.708320 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.708339 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.708354 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.810746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.810791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.810808 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.810830 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.810875 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.913761 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.913812 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.913831 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.913853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:29 crc kubenswrapper[4970]: I1124 13:47:29.913869 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:29Z","lastTransitionTime":"2025-11-24T13:47:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.017351 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.017388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.017405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.017427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.017445 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.119700 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.119755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.119771 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.119791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.119806 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.222885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.222968 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.222980 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.222999 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.223023 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.326501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.326548 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.326561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.326592 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.326603 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.429696 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.429770 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.429790 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.429821 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.429840 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.469357 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.469384 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:30 crc kubenswrapper[4970]: E1124 13:47:30.469551 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:30 crc kubenswrapper[4970]: E1124 13:47:30.469766 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.533206 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.533263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.533282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.533306 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.533322 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.636713 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.636766 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.636778 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.636795 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.636807 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.740045 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.740111 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.740134 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.740163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.740185 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.842333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.842387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.842397 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.842414 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.842425 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.945751 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.945819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.945841 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.945871 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:30 crc kubenswrapper[4970]: I1124 13:47:30.945893 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:30Z","lastTransitionTime":"2025-11-24T13:47:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.048519 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.048842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.048870 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.048896 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.048916 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.151498 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.151545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.151561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.151608 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.151625 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.254384 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.254439 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.254452 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.254466 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.254476 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.356455 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.356506 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.356524 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.356546 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.356567 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.463637 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.463706 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.463731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.463762 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.463786 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.470090 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:31 crc kubenswrapper[4970]: E1124 13:47:31.470291 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.471146 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:31 crc kubenswrapper[4970]: E1124 13:47:31.471394 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.471844 4970 scope.go:117] "RemoveContainer" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" Nov 24 13:47:31 crc kubenswrapper[4970]: E1124 13:47:31.472515 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.566867 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.566900 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.566910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.566924 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.566936 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.668525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.668622 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.668645 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.668673 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.668693 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.770158 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.770188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.770198 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.770213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.770224 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.872742 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.872838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.872865 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.872896 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.872918 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.974987 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.975059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.975082 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.975110 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:31 crc kubenswrapper[4970]: I1124 13:47:31.975128 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:31Z","lastTransitionTime":"2025-11-24T13:47:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.071646 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.071752 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.071799 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.071818 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.071834 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.091159 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.109884 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.109952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.109969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.109993 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.110011 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.139421 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.143755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.143822 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.143839 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.143862 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.143880 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.160528 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.164542 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.164592 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.164605 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.164619 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.164629 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.181459 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.184900 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.184966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.184991 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.185021 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.185047 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.198005 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:32Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.198115 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.199514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.199563 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.199614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.199645 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.199670 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.302498 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.302551 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.302563 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.302598 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.302613 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.404202 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.404263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.404282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.404305 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.404322 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.470389 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.470427 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.470569 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:32 crc kubenswrapper[4970]: E1124 13:47:32.470744 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.506364 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.506407 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.506417 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.506435 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.506447 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.609084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.609147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.609172 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.609201 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.609228 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.711642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.711685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.711698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.711715 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.711727 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.814304 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.814344 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.814354 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.814367 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.814379 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.916913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.916946 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.916954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.916965 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:32 crc kubenswrapper[4970]: I1124 13:47:32.916974 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:32Z","lastTransitionTime":"2025-11-24T13:47:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.019656 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.019701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.019710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.019724 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.019733 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.121369 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.121402 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.121410 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.121421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.121431 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.223554 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.223611 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.223622 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.223633 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.223640 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.325799 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.325842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.325851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.325865 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.325875 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.428401 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.428466 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.428478 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.428496 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.428505 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.470077 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.470135 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:33 crc kubenswrapper[4970]: E1124 13:47:33.470288 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:33 crc kubenswrapper[4970]: E1124 13:47:33.470413 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.531029 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.531087 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.531098 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.531112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.531122 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.632865 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.632912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.632928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.632943 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.632953 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.735406 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.735445 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.735452 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.735465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.735475 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.837940 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.837987 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.838000 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.838014 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.838027 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.941382 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.941446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.941465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.941493 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:33 crc kubenswrapper[4970]: I1124 13:47:33.941509 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:33Z","lastTransitionTime":"2025-11-24T13:47:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.043886 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.043929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.043938 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.043951 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.043960 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.145725 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.145762 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.145770 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.145783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.145792 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.248331 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.248378 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.248388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.248405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.248417 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.352516 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.352570 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.352598 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.352634 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.352647 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.454765 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.454803 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.454838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.454856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.454867 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.470300 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.470346 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:34 crc kubenswrapper[4970]: E1124 13:47:34.470417 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:34 crc kubenswrapper[4970]: E1124 13:47:34.470599 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.557106 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.557166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.557191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.557204 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.557215 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.658871 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.658917 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.658927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.658943 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.658955 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.761564 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.761618 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.761626 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.761640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.761650 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.864875 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.864962 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.864997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.865014 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.865025 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.967610 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.967691 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.967709 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.967731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:34 crc kubenswrapper[4970]: I1124 13:47:34.967751 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:34Z","lastTransitionTime":"2025-11-24T13:47:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.070352 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.070422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.070435 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.070451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.070463 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.173018 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.173059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.173066 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.173079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.173088 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.274928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.274984 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.274996 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.275015 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.275027 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.376734 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.376779 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.376791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.376808 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.376820 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.470106 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:35 crc kubenswrapper[4970]: E1124 13:47:35.470221 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.470335 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:35 crc kubenswrapper[4970]: E1124 13:47:35.470448 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.479214 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.479280 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.479292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.479309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.479323 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.496836 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.516282 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.535883 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.553190 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.570017 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.582077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.582116 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.582127 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.582145 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.582160 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.585723 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.600101 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.615511 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.628882 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.638831 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.650204 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.658952 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.669283 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.687682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.687736 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.687747 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.687763 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.687775 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.692378 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.702492 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.715841 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.732347 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.746067 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:35Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.789843 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.789881 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.789889 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.789902 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.789912 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.887866 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:35 crc kubenswrapper[4970]: E1124 13:47:35.888014 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:35 crc kubenswrapper[4970]: E1124 13:47:35.888226 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:07.888212187 +0000 UTC m=+103.175969480 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.891866 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.891905 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.891916 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.891933 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.891944 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.994981 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.995049 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.995068 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.995093 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:35 crc kubenswrapper[4970]: I1124 13:47:35.995111 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:35Z","lastTransitionTime":"2025-11-24T13:47:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.097719 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.097969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.098072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.098165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.098440 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.200375 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.200619 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.200703 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.200769 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.200834 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.303644 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.303977 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.304079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.304165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.304252 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.407036 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.407109 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.407126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.407151 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.407171 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.469616 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:36 crc kubenswrapper[4970]: E1124 13:47:36.469750 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.469622 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:36 crc kubenswrapper[4970]: E1124 13:47:36.469971 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.510562 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.511353 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.511430 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.511497 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.511589 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.614241 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.614732 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.614937 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.615097 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.615239 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.717668 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.717963 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.718064 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.718153 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.718243 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.821631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.821699 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.821725 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.821753 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.821775 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.923755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.924064 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.924159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.924260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:36 crc kubenswrapper[4970]: I1124 13:47:36.924353 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:36Z","lastTransitionTime":"2025-11-24T13:47:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.027230 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.027278 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.027294 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.027315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.027331 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.129239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.129289 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.129301 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.129321 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.129332 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.232068 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.232135 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.232158 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.232185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.232205 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.334452 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.334740 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.334763 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.334780 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.334793 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.437436 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.437613 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.437696 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.437780 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.437875 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.469687 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.469718 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:37 crc kubenswrapper[4970]: E1124 13:47:37.469878 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:37 crc kubenswrapper[4970]: E1124 13:47:37.470028 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.540638 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.541191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.541255 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.541315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.541378 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.643746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.643825 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.643837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.643853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.643866 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.745784 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.745842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.745860 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.745883 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.745901 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.847949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.848057 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.848075 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.848140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.848159 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.919560 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/0.log" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.919657 4970 generic.go:334] "Generic (PLEG): container finished" podID="0875a754-321a-4a3a-a798-941eefcdffe5" containerID="4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641" exitCode=1 Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.919711 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerDied","Data":"4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.920619 4970 scope.go:117] "RemoveContainer" containerID="4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.945228 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:37Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.958147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.958212 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.958229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.958254 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.958270 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:37Z","lastTransitionTime":"2025-11-24T13:47:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.962246 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:37Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.972646 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:37Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.982486 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:37Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:37 crc kubenswrapper[4970]: I1124 13:47:37.990999 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:37Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.003542 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.016513 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.028041 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.047621 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.058865 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.061790 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.061837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.061849 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.061864 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.061879 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.072873 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.089239 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.098808 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.120331 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.135379 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.149932 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.164130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.164197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.164211 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.164228 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.164239 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.169496 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.185474 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.267058 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.267113 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.267130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.267155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.267171 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.370013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.370070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.370088 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.370112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.370131 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.470057 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.470184 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:38 crc kubenswrapper[4970]: E1124 13:47:38.470322 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:38 crc kubenswrapper[4970]: E1124 13:47:38.470384 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.471855 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.471891 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.471904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.471919 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.471931 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.574952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.575013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.575030 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.575053 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.575071 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.678140 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.678196 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.678205 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.678220 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.678233 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.781378 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.781449 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.781472 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.781503 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.781526 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.883197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.883233 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.883244 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.883260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.883271 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.925484 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/0.log" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.925561 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerStarted","Data":"53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.939332 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.949372 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.969958 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.986089 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.986116 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.986124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.986137 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.986146 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:38Z","lastTransitionTime":"2025-11-24T13:47:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:38 crc kubenswrapper[4970]: I1124 13:47:38.991012 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:38Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.009364 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.032599 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.049252 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.070776 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.087302 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.089539 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.089643 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.089663 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.089686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.089702 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.105613 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.124492 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.139609 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.162342 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.177658 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.194285 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.196316 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.196359 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.196405 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.196427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.196443 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.220456 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.238137 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.258873 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:39Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.299185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.299226 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.299237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.299254 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.299268 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.402377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.402416 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.402427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.402442 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.402454 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.470333 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:39 crc kubenswrapper[4970]: E1124 13:47:39.470490 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.470333 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:39 crc kubenswrapper[4970]: E1124 13:47:39.470770 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.505013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.505057 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.505070 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.505086 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.505099 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.608038 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.608095 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.608114 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.608147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.608173 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.712542 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.712669 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.712685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.712701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.712713 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.815837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.815885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.815896 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.815911 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.815923 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.919065 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.919106 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.919119 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.919137 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:39 crc kubenswrapper[4970]: I1124 13:47:39.919151 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:39Z","lastTransitionTime":"2025-11-24T13:47:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.022637 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.022685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.022703 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.022730 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.022750 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.125906 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.126002 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.126019 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.126042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.126058 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.228868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.228922 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.228934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.228951 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.228964 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.330929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.330973 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.330985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.331001 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.331013 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.433782 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.433829 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.433841 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.433857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.433869 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.470180 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.470260 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:40 crc kubenswrapper[4970]: E1124 13:47:40.470363 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:40 crc kubenswrapper[4970]: E1124 13:47:40.470497 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.536809 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.536978 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.537010 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.537039 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.537062 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.638942 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.638969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.638977 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.638989 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.638997 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.741501 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.741560 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.741618 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.741645 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.741663 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.844222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.844301 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.844323 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.844356 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.844379 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.946858 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.946915 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.946932 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.946954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:40 crc kubenswrapper[4970]: I1124 13:47:40.946972 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:40Z","lastTransitionTime":"2025-11-24T13:47:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.049269 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.049315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.049326 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.049349 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.049361 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.151491 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.151604 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.151629 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.151660 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.151682 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.254525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.254595 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.254611 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.254631 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.254641 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.357381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.357429 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.357440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.357456 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.357468 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.460928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.461011 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.461034 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.461064 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.461088 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.470277 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:41 crc kubenswrapper[4970]: E1124 13:47:41.470454 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.470633 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:41 crc kubenswrapper[4970]: E1124 13:47:41.470843 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.563950 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.564013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.564031 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.564054 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.564072 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.666789 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.666860 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.666883 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.666911 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.666932 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.769484 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.769617 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.769637 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.769663 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.769681 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.872731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.872838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.872862 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.872892 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.872914 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.975847 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.975927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.975951 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.975984 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:41 crc kubenswrapper[4970]: I1124 13:47:41.976008 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:41Z","lastTransitionTime":"2025-11-24T13:47:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.080790 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.080876 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.080902 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.080930 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.080948 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.184117 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.184207 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.184232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.184262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.184288 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.287510 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.287559 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.287602 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.287625 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.287642 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.321928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.322021 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.322053 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.322083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.322106 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.352538 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.358282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.358347 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.358377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.358404 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.358422 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.376005 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.381044 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.381103 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.381127 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.381155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.381176 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.404635 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.409087 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.409139 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.409157 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.409180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.409198 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.430367 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.435275 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.435300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.435309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.435324 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.435335 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.456733 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:42Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.457082 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.459141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.459197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.459224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.459252 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.459274 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.469558 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.469648 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.469831 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:42 crc kubenswrapper[4970]: E1124 13:47:42.470049 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.562192 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.562248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.562264 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.562287 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.562305 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.696001 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.696072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.696096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.696124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.696145 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.799176 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.799215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.799224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.799237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.799246 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.901689 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.901782 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.901800 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.901823 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:42 crc kubenswrapper[4970]: I1124 13:47:42.901839 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:42Z","lastTransitionTime":"2025-11-24T13:47:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.004370 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.004430 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.004447 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.004470 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.004490 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.107520 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.107623 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.107642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.107666 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.107684 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.209841 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.210115 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.210272 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.210477 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.210735 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.312814 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.312856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.312864 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.312880 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.312891 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.415216 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.415260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.415274 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.415290 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.415301 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.469818 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.469867 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:43 crc kubenswrapper[4970]: E1124 13:47:43.470051 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:43 crc kubenswrapper[4970]: E1124 13:47:43.470207 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.518345 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.518398 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.518413 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.518431 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.518444 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.621740 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.621800 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.621819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.621844 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.621862 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.725813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.726695 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.726727 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.726755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.726772 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.829923 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.829991 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.830013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.830042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.830066 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.932605 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.932686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.932702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.932729 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:43 crc kubenswrapper[4970]: I1124 13:47:43.932747 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:43Z","lastTransitionTime":"2025-11-24T13:47:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.035572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.035645 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.035657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.035675 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.035687 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.137796 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.137858 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.137868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.137882 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.137891 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.241402 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.241489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.241514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.241545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.241568 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.344430 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.344494 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.344512 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.344536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.344553 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.447105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.447165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.447182 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.447205 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.447222 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.469855 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.470016 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:44 crc kubenswrapper[4970]: E1124 13:47:44.470252 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:44 crc kubenswrapper[4970]: E1124 13:47:44.470397 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.550998 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.551057 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.551073 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.551251 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.551286 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.653513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.653546 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.653557 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.653572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.653601 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.756590 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.756626 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.756635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.756649 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.756658 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.859153 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.859224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.859240 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.859263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.859275 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.962732 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.962811 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.962835 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.962865 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:44 crc kubenswrapper[4970]: I1124 13:47:44.962890 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:44Z","lastTransitionTime":"2025-11-24T13:47:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.065540 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.065665 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.065691 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.065731 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.065758 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.169520 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.169628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.169649 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.169677 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.169699 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.272390 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.272455 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.272475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.272497 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.272513 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.375251 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.375327 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.375348 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.375377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.375400 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.470332 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.470341 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:45 crc kubenswrapper[4970]: E1124 13:47:45.470523 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:45 crc kubenswrapper[4970]: E1124 13:47:45.470804 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.477695 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.477718 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.477727 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.477738 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.477746 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.487477 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.502753 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.517183 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.534608 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.549689 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.565664 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.579154 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.579215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.579236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.579260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.579279 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.594867 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.608947 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.619983 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.639770 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.653794 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.677000 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.687193 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.687232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.687242 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.687256 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.687267 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.700745 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.717333 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.732327 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.750474 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.767046 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.785777 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:45Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.789412 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.789449 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.789461 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.789478 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.789491 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.892307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.892356 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.892374 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.892396 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.892413 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.995702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.995760 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.995774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.995793 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:45 crc kubenswrapper[4970]: I1124 13:47:45.995807 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:45Z","lastTransitionTime":"2025-11-24T13:47:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.098937 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.098998 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.099016 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.099038 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.099057 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.201753 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.201815 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.201833 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.201857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.201877 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.304958 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.305004 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.305017 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.305036 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.305049 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.408061 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.408115 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.408165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.408191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.408208 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.469993 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.470012 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:46 crc kubenswrapper[4970]: E1124 13:47:46.470186 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:46 crc kubenswrapper[4970]: E1124 13:47:46.470296 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.471417 4970 scope.go:117] "RemoveContainer" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.511267 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.511305 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.511318 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.511335 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.511348 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.613749 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.614077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.614088 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.614102 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.614115 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.717440 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.717488 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.717500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.717517 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.717530 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.820371 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.820418 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.820428 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.820445 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.820455 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.923042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.923096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.923112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.923136 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.923154 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:46Z","lastTransitionTime":"2025-11-24T13:47:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.955429 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/2.log" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.959541 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.960400 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:47:46 crc kubenswrapper[4970]: I1124 13:47:46.990225 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:46Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.007462 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.025768 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.025813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.025847 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.025868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.025883 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.038712 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.062901 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.079196 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.089389 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.099457 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.111180 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.123353 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.128944 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.128978 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.128986 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.129000 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.129009 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.132032 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.149028 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.159176 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.167858 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.178619 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.191265 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.203866 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.219764 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.231062 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.231099 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.231111 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.231129 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.231140 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.234852 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.333476 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.333536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.333555 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.333600 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.333614 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.436394 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.436462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.436478 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.436502 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.436517 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.470288 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.470305 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:47 crc kubenswrapper[4970]: E1124 13:47:47.470459 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:47 crc kubenswrapper[4970]: E1124 13:47:47.470610 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.539746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.539813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.539832 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.539863 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.539882 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.643946 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.644007 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.644024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.644050 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.644067 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.746382 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.746420 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.746430 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.746444 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.746453 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.849874 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.849947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.849963 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.849985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.850004 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.953335 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.953387 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.953404 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.953426 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.953443 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:47Z","lastTransitionTime":"2025-11-24T13:47:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.966087 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.967111 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/2.log" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.971611 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" exitCode=1 Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.971653 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.971692 4970 scope.go:117] "RemoveContainer" containerID="9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.972258 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:47:47 crc kubenswrapper[4970]: E1124 13:47:47.972396 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:47 crc kubenswrapper[4970]: I1124 13:47:47.986592 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.000036 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:47Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.017746 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.031672 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.045377 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.056025 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.056096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.056116 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.056141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.056164 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.059215 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.090774 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9758d3483a5072efe642e3d615fc886606becc9de79bb16567999731288365df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:18Z\\\",\\\"message\\\":\\\"nil)\\\\nI1124 13:47:18.372279 6657 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 13:47:18.372726 6657 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.92 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {73135118-cf1b-4568-bd31-2f50308bf69d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 13:47:18.372312 6657 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76\\\\nI1124 13:47:18.372758 6657 services_controller.go:451] Built service openshift-dns/dns-default cluster-wide LB for network=default: []services.LB{}\\\\nI1124 13:47:18.372756 6657 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 13:47:18.372788 6657 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:47Z\\\",\\\"message\\\":\\\"se,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.109],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1124 13:47:47.431209 7042 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.102456 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.114009 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.128317 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.146602 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.158484 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.158552 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.158565 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.158617 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.158629 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.170431 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.186913 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.203333 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.215651 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.226983 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.243945 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.261265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.261302 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.261315 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.261328 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.261338 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.277097 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:48Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.328331 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.328475 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.328524 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.328771 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.328794 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.328920 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.328874668 +0000 UTC m=+147.616632001 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.328941 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.328992 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329034 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329051 4970 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329073 4970 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329084 4970 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.328995 4970 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329163 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.329134074 +0000 UTC m=+147.616891437 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329248 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.329228707 +0000 UTC m=+147.616986110 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329269 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.329259118 +0000 UTC m=+147.617016561 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329290 4970 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.329358 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.329345 +0000 UTC m=+147.617102443 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.363656 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.363755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.363780 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.363810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.363832 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.466811 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.466904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.466930 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.466965 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.466988 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.470276 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.470276 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.470490 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.470662 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.570975 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.571061 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.571084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.571109 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.571133 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.674383 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.674451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.674488 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.674521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.674545 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.778461 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.778519 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.778539 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.778567 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.778616 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.882124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.882193 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.882210 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.882235 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.882253 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.978226 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984341 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:47:48 crc kubenswrapper[4970]: E1124 13:47:48.984830 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984925 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984941 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984964 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:48 crc kubenswrapper[4970]: I1124 13:47:48.984981 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:48Z","lastTransitionTime":"2025-11-24T13:47:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.014118 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.039381 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.058913 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.084724 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.089176 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.089259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.089281 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.089306 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.089326 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.109408 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.127197 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.151720 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.173466 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.191946 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.191990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.192005 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.192025 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.192042 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.193335 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.209546 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.221823 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.234088 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.246837 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.263134 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.279051 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.295433 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.295686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.295857 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.295992 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.296135 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.298819 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.330470 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:47Z\\\",\\\"message\\\":\\\"se,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.109],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1124 13:47:47.431209 7042 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.348017 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:49Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.400046 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.400465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.400635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.400819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.400955 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.469916 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:49 crc kubenswrapper[4970]: E1124 13:47:49.470130 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.470671 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:49 crc kubenswrapper[4970]: E1124 13:47:49.471094 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.505004 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.505068 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.505085 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.505108 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.505129 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.607779 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.607851 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.607872 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.607900 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.607919 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.711032 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.711079 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.711114 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.711149 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.711170 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.814507 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.814627 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.814653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.814684 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.814705 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.917247 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.917337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.917356 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.917381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:49 crc kubenswrapper[4970]: I1124 13:47:49.917400 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:49Z","lastTransitionTime":"2025-11-24T13:47:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.020218 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.020263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.020274 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.020290 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.020302 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.123194 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.123259 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.123277 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.123302 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.123320 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.227045 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.227113 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.227131 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.227155 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.227173 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.331150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.331215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.331237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.331265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.331287 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.469559 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.469606 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:50 crc kubenswrapper[4970]: E1124 13:47:50.469817 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:50 crc kubenswrapper[4970]: E1124 13:47:50.469951 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.471607 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.471651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.471692 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.471709 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.471721 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.574462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.574508 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.574525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.574547 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.574562 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.677485 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.677525 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.677536 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.677548 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.677557 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.781966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.782378 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.782510 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.782677 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.782810 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.885615 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.885665 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.885680 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.885698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.885713 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.988977 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.989040 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.989127 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.989160 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:50 crc kubenswrapper[4970]: I1124 13:47:50.989179 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:50Z","lastTransitionTime":"2025-11-24T13:47:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.092048 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.092111 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.092128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.092151 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.092169 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.195853 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.195949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.195982 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.196014 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.196037 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.299358 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.299423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.299442 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.299465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.299484 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.402824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.402910 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.402941 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.402970 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.402988 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.469659 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.469714 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:51 crc kubenswrapper[4970]: E1124 13:47:51.469873 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:51 crc kubenswrapper[4970]: E1124 13:47:51.470003 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.505892 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.505952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.505971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.505995 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.506012 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.609062 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.609166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.609187 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.609215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.609232 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.712085 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.712144 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.712166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.712189 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.712207 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.815524 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.815650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.815672 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.815730 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.815749 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.918520 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.918632 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.918657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.918686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:51 crc kubenswrapper[4970]: I1124 13:47:51.918709 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:51Z","lastTransitionTime":"2025-11-24T13:47:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.022929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.023085 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.023152 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.023180 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.023241 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.129940 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.130021 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.130040 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.130067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.130094 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.232860 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.232922 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.232939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.232964 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.232981 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.335460 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.335545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.335572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.335638 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.335711 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.439039 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.439084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.439095 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.439111 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.439121 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.463399 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.463464 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.463487 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.463518 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.463540 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.471870 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.471944 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.471989 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.472137 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.482719 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.487224 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.487294 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.487322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.487353 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.487376 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.506945 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.511993 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.512023 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.512031 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.512046 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.512055 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.524669 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.529126 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.529191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.529212 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.529243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.529265 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.547753 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.552885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.552934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.552953 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.552977 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.552996 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.572860 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:52Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:52 crc kubenswrapper[4970]: E1124 13:47:52.573196 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.575686 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.575747 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.575769 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.575793 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.575811 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.679257 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.679314 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.679334 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.679359 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.679377 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.781819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.782250 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.782281 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.782313 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.782335 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.885016 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.885080 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.885099 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.885124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.885141 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.988105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.988169 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.988197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.988226 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:52 crc kubenswrapper[4970]: I1124 13:47:52.988244 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:52Z","lastTransitionTime":"2025-11-24T13:47:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.090847 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.090907 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.090923 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.090944 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.090961 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.194310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.194377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.194398 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.194466 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.194491 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.297377 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.297427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.297438 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.297456 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.297468 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.399751 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.399822 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.399856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.399887 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.399909 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.469891 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.469892 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:53 crc kubenswrapper[4970]: E1124 13:47:53.470130 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:53 crc kubenswrapper[4970]: E1124 13:47:53.470251 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.503142 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.503213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.503232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.503258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.503279 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.606360 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.606465 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.606515 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.606549 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.606569 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.710030 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.710143 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.710170 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.710199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.710222 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.814651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.814729 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.814753 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.814783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.814804 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.918047 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.918128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.918153 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.918185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:53 crc kubenswrapper[4970]: I1124 13:47:53.918212 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:53Z","lastTransitionTime":"2025-11-24T13:47:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.022179 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.022245 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.022262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.022287 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.022305 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.129299 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.129368 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.129388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.129416 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.129436 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.232489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.232535 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.232548 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.232565 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.232593 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.335756 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.335813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.335837 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.335864 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.335885 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.439500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.439572 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.439623 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.439653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.439675 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.470045 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:54 crc kubenswrapper[4970]: E1124 13:47:54.470212 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.470453 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:54 crc kubenswrapper[4970]: E1124 13:47:54.470553 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.542621 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.542677 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.542695 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.542718 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.542737 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.645983 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.646054 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.646093 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.646123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.646145 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.749621 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.749697 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.749715 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.749741 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.749759 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.852785 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.852839 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.852856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.852881 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.852899 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.956112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.956188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.956207 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.956230 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:54 crc kubenswrapper[4970]: I1124 13:47:54.956249 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:54Z","lastTransitionTime":"2025-11-24T13:47:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.059280 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.059332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.059343 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.059358 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.059369 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.161327 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.161412 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.161446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.161474 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.161494 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.263484 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.263545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.263563 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.263614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.263634 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.366565 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.366657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.366678 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.366705 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.366729 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469355 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469374 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469437 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.469517 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:55 crc kubenswrapper[4970]: E1124 13:47:55.470086 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.470138 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: E1124 13:47:55.470322 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.494909 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.516318 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.533669 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.551663 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.573654 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.573764 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.573788 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.573817 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.573844 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.575100 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1656075f-8f28-451a-8d60-c49e32fa593d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8af0006096f06c4219c2cb7117eb43fffb29dfd81fca144231027ca39abe0041\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbfea39b1c5b85ab008b0472179c77434645b8cbbbae97c3b90c5005c8fc422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38657e8a4c161a76340afe0e7343e13b9255dfd9215b4cb287994e632b299996\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d99188ddd5d94952f7020b43cc2cf47edba3d8594f52cb85df12b7e1fb5e4377\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.598964 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2da682ae5e74bb04045a69be9a869ac2bc44cbf746db68928d9aaf6f7ece63ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.612021 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.639623 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b38f92e-6c99-47b9-9079-0b0e455ddca1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:47Z\\\",\\\"message\\\":\\\"se,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.109],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1124 13:47:47.431209 7042 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:47:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r754h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-msd5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.656520 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tc8pw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2bfd2947-8ebd-4cfa-b412-eec444657731\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28a0a191f87d9d6d8e48fa66b1e4c1e7d784942b14a69a96a8bde2d5181cd608\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6jpxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tc8pw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.677319 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.677370 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.677388 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.677409 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.677424 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.680649 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b0fceff-420b-4465-9afa-99af03170306\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kjt8w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kv7zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.701050 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-vtfnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0875a754-321a-4a3a-a798-941eefcdffe5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T13:47:37Z\\\",\\\"message\\\":\\\"2025-11-24T13:46:51+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b\\\\n2025-11-24T13:46:51+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1cb95a2-b30e-4943-a08c-e31128a2132b to /host/opt/cni/bin/\\\\n2025-11-24T13:46:52Z [verbose] multus-daemon started\\\\n2025-11-24T13:46:52Z [verbose] Readiness Indicator file check\\\\n2025-11-24T13:47:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cn7wj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-vtfnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.724015 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64b0c636-eda5-4094-b61e-26f6885f431b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5df0d2c79180b7725669f6185aab73423b84e2ad73d7ce0633313c9347a28f7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://944483fdb9b971699b034e59fd95b846a2797d384120f0fd126464e0dc2d38cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://598749d408cfbc6eb60b41b2d6a49a0d95b440bc2a5cba386ce56bb8b523c8dc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8b0ac4b138100b277f3cd150f5b690bc3e109a73f62e0fffbc45dd725e163\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b456edc708c04ec179ab0fc516ae20859d434687dddd0f553935b8946b447258\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a73c81c1188978a4c8c956b9f1cd09418beae2a3f9f44445527bcdf696733d61\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eeabc390b2bb17d13837fd07012009171ed2a746d739a0ecdc755c78d103ee2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6kmkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-gr7xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.743106 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8aac2e9-6f3d-4453-bf06-fc4ef1647cb5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b99744d665f397ad2760a9c69908261d905a850980c1846f4be27d1b32f7af8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ba02dbb5243e74ec69f622a11a40bc6edffc14d8c79484e662cdbb63e9d4c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f02199603ac9012b18d7ce25566234061704c2c2db5f65fe9cb1834d6c4f5f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f1f07608d1f07b97fb7b85fb2681aeceff76a9aa837a3963eae8f60ca02ea3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://359e613d5f3c3314c90edbdd6c686c0b75c367599922f4400022955e3764fafc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7150e9276f8c612b732413eb39ccdbe56752a11047b5eba66737d0994499aecf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://810e25a3c3a0b007173897c6f16f03885dfea29e76254e2ee8e5aa5f876bcc29\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd441e6ba37e6a25d1136175181d450e3b39b43fd2a8ec5aadbf4b8dc96b6749\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.761384 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57d92547-21c3-4d83-970b-933ec88c7ff6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7e8a49b956f192a2009d496e5e6ab78eaa78188842fd8fcdaf4f1c39382a6e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://735091de4a613114d8d7904b6d858feb440d167a9db10cceb10f918492b49f79\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e80c340f0dfcbfc53e25c54d42416190d19ab1841e6c4ffcc7af46dbfaaa8fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3318f3c1e6836173ae652ca24fa092af26b3aa49ea97e7e53cd9e2053403ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f472b70b893402e1239ba8e89f01b4e3cc44c7dd35902ca3299ff74584fb1cfa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4172c79615219a589303290f8ce930f12af8da4ef77de41bd294445c890ac5c8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.775162 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28081a08-4f1d-4e72-b269-ef1cabf10c61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0748870d5f29b4a35a43599d825e134f57f46622f82db090228e6ed75b9e2e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://823ba1295f227a114ecc2bfda298bdef622dc49794534be612252484c17ad3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ef4df937eb6b9e8501d7a0b6cf5d1c2cf189e01b83d08d43e6db09282c9014f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75b7e35db8a348dd974bdd263006bf035c7d45dd2d290b29b9ab219b8874d95f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T13:46:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T13:46:26Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.779275 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.779310 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.779322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.779337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.779349 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.791235 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51eedf686e01f11486932d3be3246f87131836e35ea0774dc120b37c91111fcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.809862 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.826217 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"642b8bf5-104e-4448-9f11-8c2dd5d12c41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:47:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://810d11500d044c1823ebb8bc8b108449f93525650acb8a9beb28351820a2d5d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fb0a0e6cf076208bb99a3c5fe3f80cc8d61a4bf01346980794e0699a14e90e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:47:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h98nv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:47:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4nq76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:47:55Z is after 2025-08-24T17:21:41Z" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.881591 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.881628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.881639 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.881656 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.881666 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.983990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.984062 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.984098 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.984128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:55 crc kubenswrapper[4970]: I1124 13:47:55.984150 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:55Z","lastTransitionTime":"2025-11-24T13:47:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.086852 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.086909 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.086926 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.086949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.086967 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.190443 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.190485 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.190497 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.190513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.190524 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.293359 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.293413 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.293427 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.293445 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.293457 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.396359 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.396414 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.396433 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.396460 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.396478 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.469490 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.469490 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:56 crc kubenswrapper[4970]: E1124 13:47:56.469715 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:56 crc kubenswrapper[4970]: E1124 13:47:56.469867 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.484829 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.499783 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.499942 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.499973 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.500004 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.500047 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.603955 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.604067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.604114 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.604146 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.604167 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.707691 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.707767 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.707795 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.707828 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.707854 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.810935 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.811050 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.811122 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.811148 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.811303 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.913889 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.913949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.913966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.913993 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:56 crc kubenswrapper[4970]: I1124 13:47:56.914017 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:56Z","lastTransitionTime":"2025-11-24T13:47:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.017169 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.017258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.017290 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.017320 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.017344 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.120253 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.120319 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.120341 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.120371 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.120394 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.222996 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.223055 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.223072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.223096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.223126 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.325648 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.325733 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.325745 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.325757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.325766 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.428478 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.428549 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.428568 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.428620 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.428639 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.470488 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.470488 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:57 crc kubenswrapper[4970]: E1124 13:47:57.470835 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:57 crc kubenswrapper[4970]: E1124 13:47:57.470937 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.531435 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.531499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.531520 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.531549 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.531688 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.635990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.636076 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.636096 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.636123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.636140 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.738897 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.739267 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.739424 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.739637 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.739819 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.843642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.843716 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.843741 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.843774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.843796 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.947017 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.947073 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.947089 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.947108 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:57 crc kubenswrapper[4970]: I1124 13:47:57.947122 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:57Z","lastTransitionTime":"2025-11-24T13:47:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.050962 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.051277 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.051444 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.051685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.051870 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.155192 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.155637 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.155846 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.155992 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.156139 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.259947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.259997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.260014 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.260041 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.260058 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.363671 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.363762 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.363785 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.363816 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.363838 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.466385 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.466481 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.466499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.466557 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.466606 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.470125 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.470164 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:47:58 crc kubenswrapper[4970]: E1124 13:47:58.470276 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:47:58 crc kubenswrapper[4970]: E1124 13:47:58.470417 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.569367 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.569433 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.569451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.569474 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.569492 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.672724 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.672777 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.672795 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.672819 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.672836 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.774984 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.775455 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.775741 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.776067 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.776245 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.879878 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.880008 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.880034 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.880058 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.880074 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.983083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.983189 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.983261 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.983302 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:58 crc kubenswrapper[4970]: I1124 13:47:58.983325 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:58Z","lastTransitionTime":"2025-11-24T13:47:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.085810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.085879 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.085901 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.085929 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.085952 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.188952 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.189011 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.189030 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.189053 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.189075 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.291912 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.291997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.292023 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.292052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.292080 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.395062 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.395138 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.395162 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.395190 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.395211 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.469956 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.469968 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:47:59 crc kubenswrapper[4970]: E1124 13:47:59.470414 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:47:59 crc kubenswrapper[4970]: E1124 13:47:59.470569 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.498118 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.498225 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.498284 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.498312 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.498340 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.601022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.601112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.601135 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.601168 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.601190 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.703871 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.703943 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.703967 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.703996 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.704018 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.807143 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.807194 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.807204 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.807218 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.807228 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.910060 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.910118 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.910134 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.910236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:47:59 crc kubenswrapper[4970]: I1124 13:47:59.910260 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:47:59Z","lastTransitionTime":"2025-11-24T13:47:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.013203 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.013280 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.013300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.013323 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.013340 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.115934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.115993 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.116005 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.116019 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.116029 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.218641 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.218684 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.218693 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.218706 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.218715 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.322754 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.322823 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.322842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.322867 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.322884 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.426130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.426192 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.426209 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.426239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.426256 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.470123 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.470280 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:00 crc kubenswrapper[4970]: E1124 13:48:00.470513 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:00 crc kubenswrapper[4970]: E1124 13:48:00.471021 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.529059 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.529119 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.529135 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.529159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.529177 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.632144 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.632200 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.632212 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.632229 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.632241 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.734436 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.734835 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.734990 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.735141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.735266 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.837532 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.837809 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.837876 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.837948 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.838021 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.940514 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.940833 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.940911 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.940981 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:00 crc kubenswrapper[4970]: I1124 13:48:00.941045 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:00Z","lastTransitionTime":"2025-11-24T13:48:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.043685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.043726 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.043739 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.043757 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.043769 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.147004 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.147423 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.147640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.147808 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.147997 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.251389 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.251435 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.251463 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.251486 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.251501 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.354561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.354682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.354710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.354744 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.354768 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.458232 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.458617 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.458635 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.458658 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.458675 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.469697 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.469706 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:01 crc kubenswrapper[4970]: E1124 13:48:01.469964 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:01 crc kubenswrapper[4970]: E1124 13:48:01.470049 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.562513 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.562607 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.562627 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.562650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.562667 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.665939 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.666022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.666044 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.666075 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.666098 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.769207 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.769257 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.769271 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.769292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.769307 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.872056 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.872178 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.872203 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.872242 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.872265 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.974969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.975011 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.975022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.975072 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:01 crc kubenswrapper[4970]: I1124 13:48:01.975095 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:01Z","lastTransitionTime":"2025-11-24T13:48:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.077547 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.077627 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.077642 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.077661 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.077676 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.180317 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.180475 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.180493 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.180518 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.180534 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.283934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.284022 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.284048 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.284077 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.284100 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.387204 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.387265 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.387282 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.387306 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.387322 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.470112 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:02 crc kubenswrapper[4970]: E1124 13:48:02.474560 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.475996 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:02 crc kubenswrapper[4970]: E1124 13:48:02.476178 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.490407 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.490479 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.490499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.490524 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.490541 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.593672 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.593721 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.593732 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.593749 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.593760 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.696532 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.696615 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.696640 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.696660 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.696673 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.798614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.798656 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.798668 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.798683 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.798694 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.902168 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.902222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.902241 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.902264 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.902283 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.935541 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.935651 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.935701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.935730 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.935751 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: E1124 13:48:02.958080 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.963843 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.963914 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.963938 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.963969 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.963992 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:02 crc kubenswrapper[4970]: E1124 13:48:02.989187 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:02Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.995142 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.995206 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.995228 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.995258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:02 crc kubenswrapper[4970]: I1124 13:48:02.995281 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:02Z","lastTransitionTime":"2025-11-24T13:48:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.018276 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.023499 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.023534 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.023545 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.023562 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.023593 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.044096 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.048697 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.048869 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.048894 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.048973 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.049001 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.068870 4970 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T13:48:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"40be1732-2104-461a-99c4-0be47e2bfcda\\\",\\\"systemUUID\\\":\\\"e4e3f2ab-097f-4c21-9d01-3cfa8ecdc590\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:03Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.069088 4970 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.071249 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.071307 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.071322 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.071342 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.071356 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.176379 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.176444 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.176462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.176482 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.176495 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.279186 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.279236 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.279248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.279266 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.279279 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.382820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.382885 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.382902 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.382927 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.382945 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.470221 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.470566 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.470706 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.470804 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.470859 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:03 crc kubenswrapper[4970]: E1124 13:48:03.470943 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.485975 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.486032 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.486049 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.486071 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.486088 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.588763 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.588813 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.588824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.588842 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.588852 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.692266 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.692366 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.692384 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.692445 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.692464 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.795614 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.795675 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.795693 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.795721 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.795739 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.898548 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.898654 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.898682 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.898710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:03 crc kubenswrapper[4970]: I1124 13:48:03.898733 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:03Z","lastTransitionTime":"2025-11-24T13:48:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.002376 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.002462 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.002486 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.002521 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.002544 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.104917 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.104984 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.105006 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.105039 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.105064 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.207657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.207718 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.207740 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.207770 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.207792 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.310195 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.310252 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.310271 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.310301 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.310318 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.413884 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.413954 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.413971 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.413998 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.414022 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.469741 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.469740 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:04 crc kubenswrapper[4970]: E1124 13:48:04.469906 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:04 crc kubenswrapper[4970]: E1124 13:48:04.470012 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.516674 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.516737 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.516791 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.516811 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.516829 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.619163 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.619351 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.619392 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.619421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.619444 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.722362 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.722422 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.722443 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.722478 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.722502 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.824866 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.824909 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.824917 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.824930 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.824939 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.927668 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.927701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.927710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.927721 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:04 crc kubenswrapper[4970]: I1124 13:48:04.927729 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:04Z","lastTransitionTime":"2025-11-24T13:48:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.031192 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.031262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.031279 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.031303 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.031319 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.134839 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.134913 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.134934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.134959 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.134978 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.237960 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.238042 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.238064 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.238088 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.238136 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.341643 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.341707 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.341723 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.341746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.341762 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.445193 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.445250 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.445270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.445293 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.445311 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.469890 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:05 crc kubenswrapper[4970]: E1124 13:48:05.470441 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.471047 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:05 crc kubenswrapper[4970]: E1124 13:48:05.471269 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.488073 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5dd7b91d-1aca-41aa-b8b4-ab97723e8074\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6eb66fb42602232a3db7a2c8ee7ea5c33941b632e3fb700bbbdb161228e9b352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r66kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9jtl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.510397 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18a12e804155ab7fca371f300d49d329fed934ee18e42b0a955cfd183b8b741a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e93f472061aff8d682aa845d3f50201f8fdd43b70cf03cd3e743224fc46333d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.529631 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:44Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.546690 4970 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bhjg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2628d9f-c695-4d55-bf97-30c4443ca0d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T13:46:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e4b93eec48c4290aed1792b922c5385ca2dfe283ea27edf6c7c1c2adb3e689c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T13:46:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xmfvl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T13:46:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bhjg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T13:48:05Z is after 2025-08-24T17:21:41Z" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.549189 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.549243 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.549263 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.549563 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.549666 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.616393 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tc8pw" podStartSLOduration=76.616359856 podStartE2EDuration="1m16.616359856s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.615857372 +0000 UTC m=+100.903614705" watchObservedRunningTime="2025-11-24 13:48:05.616359856 +0000 UTC m=+100.904117199" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.651768 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.651810 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.651824 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.652866 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=77.652841627 podStartE2EDuration="1m17.652841627s" podCreationTimestamp="2025-11-24 13:46:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.652646301 +0000 UTC m=+100.940403675" watchObservedRunningTime="2025-11-24 13:48:05.652841627 +0000 UTC m=+100.940598940" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.653120 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.653146 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.722675 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=9.722646605 podStartE2EDuration="9.722646605s" podCreationTimestamp="2025-11-24 13:47:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.705784391 +0000 UTC m=+100.993541744" watchObservedRunningTime="2025-11-24 13:48:05.722646605 +0000 UTC m=+101.010403978" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.723223 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-vtfnd" podStartSLOduration=76.72321181 podStartE2EDuration="1m16.72321181s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.723155458 +0000 UTC m=+101.010912741" watchObservedRunningTime="2025-11-24 13:48:05.72321181 +0000 UTC m=+101.010969143" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.767888 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-gr7xg" podStartSLOduration=76.767861616 podStartE2EDuration="1m16.767861616s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.760549334 +0000 UTC m=+101.048306637" watchObservedRunningTime="2025-11-24 13:48:05.767861616 +0000 UTC m=+101.055618949" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.774928 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.774985 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.775002 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.775024 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.775040 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.870916 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4nq76" podStartSLOduration=76.87089046 podStartE2EDuration="1m16.87089046s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.845875851 +0000 UTC m=+101.133633154" watchObservedRunningTime="2025-11-24 13:48:05.87089046 +0000 UTC m=+101.158647753" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.871062 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=81.871057875 podStartE2EDuration="1m21.871057875s" podCreationTimestamp="2025-11-24 13:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.869038981 +0000 UTC m=+101.156796304" watchObservedRunningTime="2025-11-24 13:48:05.871057875 +0000 UTC m=+101.158815168" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.877762 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.877821 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.877902 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.877973 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.877993 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.884566 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.884550549 podStartE2EDuration="1m21.884550549s" podCreationTimestamp="2025-11-24 13:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.884390105 +0000 UTC m=+101.172147438" watchObservedRunningTime="2025-11-24 13:48:05.884550549 +0000 UTC m=+101.172307842" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.899956 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.899938315 podStartE2EDuration="46.899938315s" podCreationTimestamp="2025-11-24 13:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:05.899206506 +0000 UTC m=+101.186963839" watchObservedRunningTime="2025-11-24 13:48:05.899938315 +0000 UTC m=+101.187695618" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.981117 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.981159 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.981171 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.981188 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:05 crc kubenswrapper[4970]: I1124 13:48:05.981198 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:05Z","lastTransitionTime":"2025-11-24T13:48:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.083348 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.083404 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.083421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.083446 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.083464 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.186018 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.186083 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.186099 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.186124 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.186141 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.289764 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.289825 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.289843 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.289868 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.289885 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.393218 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.393312 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.393337 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.393367 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.393390 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.470380 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.470445 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:06 crc kubenswrapper[4970]: E1124 13:48:06.470838 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:06 crc kubenswrapper[4970]: E1124 13:48:06.471077 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.496106 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.496189 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.496213 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.496239 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.496258 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.600540 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.600636 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.600657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.600683 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.600700 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.704357 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.704418 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.704443 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.704473 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.704495 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.807242 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.807287 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.807300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.807317 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.807328 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.910966 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.911028 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.911046 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.911075 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:06 crc kubenswrapper[4970]: I1124 13:48:06.911098 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:06Z","lastTransitionTime":"2025-11-24T13:48:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.013069 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.013142 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.013168 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.013197 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.013219 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.116170 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.116237 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.116248 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.116283 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.116296 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.218479 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.218535 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.218546 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.218565 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.218599 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.320973 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.321025 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.321035 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.321052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.321063 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.423262 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.423330 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.423346 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.423376 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.423394 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.469823 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.469877 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:07 crc kubenswrapper[4970]: E1124 13:48:07.470009 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:07 crc kubenswrapper[4970]: E1124 13:48:07.470122 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.526843 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.526904 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.526926 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.526947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.526962 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.630623 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.630693 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.630713 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.630736 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.630754 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.733776 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.733821 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.733833 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.733850 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.733861 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.836381 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.836433 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.836459 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.836484 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.836499 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.938742 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.938777 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.938785 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.938797 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.938805 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:07Z","lastTransitionTime":"2025-11-24T13:48:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:07 crc kubenswrapper[4970]: I1124 13:48:07.972491 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:07 crc kubenswrapper[4970]: E1124 13:48:07.972649 4970 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:48:07 crc kubenswrapper[4970]: E1124 13:48:07.972699 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs podName:8b0fceff-420b-4465-9afa-99af03170306 nodeName:}" failed. No retries permitted until 2025-11-24 13:49:11.972685082 +0000 UTC m=+167.260442375 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs") pod "network-metrics-daemon-kv7zv" (UID: "8b0fceff-420b-4465-9afa-99af03170306") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.041038 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.041080 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.041090 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.041105 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.041117 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.143717 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.143798 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.143838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.143871 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.143897 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.247222 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.247301 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.247342 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.247372 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.247394 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.350980 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.351058 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.351076 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.351107 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.351125 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.453670 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.453745 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.453769 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.453798 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.453820 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.470141 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.470291 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:08 crc kubenswrapper[4970]: E1124 13:48:08.470723 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:08 crc kubenswrapper[4970]: E1124 13:48:08.470808 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.559907 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.559949 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.559958 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.559976 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.559985 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.662685 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.662756 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.662771 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.662794 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.662809 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.767169 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.767240 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.767258 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.767286 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.767305 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.870057 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.870108 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.870119 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.870137 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.870149 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.973068 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.973130 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.973146 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.973165 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:08 crc kubenswrapper[4970]: I1124 13:48:08.973179 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:08Z","lastTransitionTime":"2025-11-24T13:48:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.075901 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.075997 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.076056 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.076088 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.076109 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.179099 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.179151 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.179166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.179185 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.179200 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.282370 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.282447 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.282468 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.282492 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.282510 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.385489 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.385620 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.385650 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.385698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.385723 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.469895 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:09 crc kubenswrapper[4970]: E1124 13:48:09.470094 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.470164 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:09 crc kubenswrapper[4970]: E1124 13:48:09.470550 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.488547 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.488647 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.488666 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.488690 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.488708 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.592052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.592123 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.592141 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.592166 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.592183 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.694644 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.694710 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.694728 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.694753 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.694770 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.797993 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.798065 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.798085 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.798109 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.798126 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.901350 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.901437 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.901469 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.901500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:09 crc kubenswrapper[4970]: I1124 13:48:09.901522 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:09Z","lastTransitionTime":"2025-11-24T13:48:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.004956 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.005050 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.005081 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.005112 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.005135 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.109001 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.109053 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.109065 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.109084 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.109097 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.212249 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.212314 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.212332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.212357 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.212374 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.315935 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.316023 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.316052 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.316078 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.316096 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.419235 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.419292 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.419309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.419332 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.419349 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.470052 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.470103 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:10 crc kubenswrapper[4970]: E1124 13:48:10.470259 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:10 crc kubenswrapper[4970]: E1124 13:48:10.470631 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.522365 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.522442 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.522460 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.522486 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.522509 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.625526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.625633 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.625653 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.625680 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.625699 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.729076 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.729131 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.729147 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.729170 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.729187 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.832148 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.832199 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.832215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.832238 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.832254 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.936300 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.936366 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.936390 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.936421 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:10 crc kubenswrapper[4970]: I1124 13:48:10.936442 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:10Z","lastTransitionTime":"2025-11-24T13:48:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.040061 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.040128 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.040150 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.040181 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.040201 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.143689 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.143756 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.143774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.143799 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.143819 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.247524 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.247630 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.247657 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.247708 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.247732 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.350753 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.350820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.350838 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.350856 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.350869 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.454500 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.454629 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.454658 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.454691 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.454712 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.470280 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.470330 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:11 crc kubenswrapper[4970]: E1124 13:48:11.470446 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:11 crc kubenswrapper[4970]: E1124 13:48:11.470624 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.557408 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.557487 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.557503 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.557526 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.557545 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.660167 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.660253 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.660279 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.660309 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.660326 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.763561 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.763675 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.763694 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.763758 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.763785 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.867131 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.867191 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.867238 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.867270 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.867296 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.971215 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.971273 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.971293 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.971316 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:11 crc kubenswrapper[4970]: I1124 13:48:11.971335 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:11Z","lastTransitionTime":"2025-11-24T13:48:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.073628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.073702 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.073726 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.073755 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.073779 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.177221 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.177302 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.177333 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.177363 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.177386 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.280260 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.280327 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.280445 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.280487 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.280511 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.383389 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.383451 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.383471 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.383494 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.383513 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.469371 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.469393 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:12 crc kubenswrapper[4970]: E1124 13:48:12.469667 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:12 crc kubenswrapper[4970]: E1124 13:48:12.469840 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.486693 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.486734 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.486745 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.486764 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.486777 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.590013 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.590078 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.590101 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.590131 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.590154 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.692768 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.692875 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.692934 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.692960 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.693014 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.795882 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.795963 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.795974 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.795989 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.796004 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.899606 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.899664 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.899677 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.899698 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:12 crc kubenswrapper[4970]: I1124 13:48:12.899714 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:12Z","lastTransitionTime":"2025-11-24T13:48:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.002884 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.002947 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.002959 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.002976 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.002989 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:13Z","lastTransitionTime":"2025-11-24T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.104644 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.104705 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.104722 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.104746 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.104763 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:13Z","lastTransitionTime":"2025-11-24T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.207706 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.207774 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.207792 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.207820 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.207838 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:13Z","lastTransitionTime":"2025-11-24T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.213534 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.213628 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.213701 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.213726 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.213748 4970 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T13:48:13Z","lastTransitionTime":"2025-11-24T13:48:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.278000 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6"] Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.278522 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.281515 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.281680 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.281795 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.281857 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.355304 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bhjg8" podStartSLOduration=84.355269082 podStartE2EDuration="1m24.355269082s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:13.353728111 +0000 UTC m=+108.641485434" watchObservedRunningTime="2025-11-24 13:48:13.355269082 +0000 UTC m=+108.643026415" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.369977 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podStartSLOduration=84.369952818 podStartE2EDuration="1m24.369952818s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:13.368987543 +0000 UTC m=+108.656744866" watchObservedRunningTime="2025-11-24 13:48:13.369952818 +0000 UTC m=+108.657710121" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.439725 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.439829 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da40fac8-9c31-4ddc-9af8-72fc49b95bac-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.439877 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.439914 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da40fac8-9c31-4ddc-9af8-72fc49b95bac-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.439939 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da40fac8-9c31-4ddc-9af8-72fc49b95bac-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.470177 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.470337 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:13 crc kubenswrapper[4970]: E1124 13:48:13.470509 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:13 crc kubenswrapper[4970]: E1124 13:48:13.470668 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541456 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da40fac8-9c31-4ddc-9af8-72fc49b95bac-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541539 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541658 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da40fac8-9c31-4ddc-9af8-72fc49b95bac-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541712 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da40fac8-9c31-4ddc-9af8-72fc49b95bac-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541788 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541816 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.541896 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da40fac8-9c31-4ddc-9af8-72fc49b95bac-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.543111 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da40fac8-9c31-4ddc-9af8-72fc49b95bac-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.550867 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da40fac8-9c31-4ddc-9af8-72fc49b95bac-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.569933 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da40fac8-9c31-4ddc-9af8-72fc49b95bac-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5wbs6\" (UID: \"da40fac8-9c31-4ddc-9af8-72fc49b95bac\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:13 crc kubenswrapper[4970]: I1124 13:48:13.604145 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" Nov 24 13:48:14 crc kubenswrapper[4970]: I1124 13:48:14.080286 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" event={"ID":"da40fac8-9c31-4ddc-9af8-72fc49b95bac","Type":"ContainerStarted","Data":"1ed0b54be26708bc883f960bfaf1a90a79eb14dcea7f660405e7c5133165b3dc"} Nov 24 13:48:14 crc kubenswrapper[4970]: I1124 13:48:14.080700 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" event={"ID":"da40fac8-9c31-4ddc-9af8-72fc49b95bac","Type":"ContainerStarted","Data":"46c0c1db2821f61819739ead828e492fa32525a5be87d8f4160051d20f50235b"} Nov 24 13:48:14 crc kubenswrapper[4970]: I1124 13:48:14.469511 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:14 crc kubenswrapper[4970]: I1124 13:48:14.469646 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:14 crc kubenswrapper[4970]: E1124 13:48:14.469682 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:14 crc kubenswrapper[4970]: E1124 13:48:14.469920 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:15 crc kubenswrapper[4970]: I1124 13:48:15.469728 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:15 crc kubenswrapper[4970]: I1124 13:48:15.469728 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:15 crc kubenswrapper[4970]: E1124 13:48:15.471660 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:15 crc kubenswrapper[4970]: E1124 13:48:15.471793 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:16 crc kubenswrapper[4970]: I1124 13:48:16.469760 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:16 crc kubenswrapper[4970]: I1124 13:48:16.469804 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:16 crc kubenswrapper[4970]: E1124 13:48:16.469918 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:16 crc kubenswrapper[4970]: E1124 13:48:16.469991 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:16 crc kubenswrapper[4970]: I1124 13:48:16.470936 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:48:16 crc kubenswrapper[4970]: E1124 13:48:16.471200 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-msd5k_openshift-ovn-kubernetes(7b38f92e-6c99-47b9-9079-0b0e455ddca1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" Nov 24 13:48:17 crc kubenswrapper[4970]: I1124 13:48:17.471792 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:17 crc kubenswrapper[4970]: I1124 13:48:17.472660 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:17 crc kubenswrapper[4970]: E1124 13:48:17.472799 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:17 crc kubenswrapper[4970]: E1124 13:48:17.473014 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:18 crc kubenswrapper[4970]: I1124 13:48:18.470193 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:18 crc kubenswrapper[4970]: I1124 13:48:18.470242 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:18 crc kubenswrapper[4970]: E1124 13:48:18.470314 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:18 crc kubenswrapper[4970]: E1124 13:48:18.470369 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:19 crc kubenswrapper[4970]: I1124 13:48:19.469673 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:19 crc kubenswrapper[4970]: I1124 13:48:19.469808 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:19 crc kubenswrapper[4970]: E1124 13:48:19.469856 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:19 crc kubenswrapper[4970]: E1124 13:48:19.470048 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:20 crc kubenswrapper[4970]: I1124 13:48:20.469662 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:20 crc kubenswrapper[4970]: I1124 13:48:20.469667 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:20 crc kubenswrapper[4970]: E1124 13:48:20.469853 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:20 crc kubenswrapper[4970]: E1124 13:48:20.469960 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:21 crc kubenswrapper[4970]: I1124 13:48:21.469835 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:21 crc kubenswrapper[4970]: E1124 13:48:21.470025 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:21 crc kubenswrapper[4970]: I1124 13:48:21.470118 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:21 crc kubenswrapper[4970]: E1124 13:48:21.470312 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:22 crc kubenswrapper[4970]: I1124 13:48:22.469940 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:22 crc kubenswrapper[4970]: I1124 13:48:22.470134 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:22 crc kubenswrapper[4970]: E1124 13:48:22.470162 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:22 crc kubenswrapper[4970]: E1124 13:48:22.470293 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:23 crc kubenswrapper[4970]: I1124 13:48:23.470146 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:23 crc kubenswrapper[4970]: E1124 13:48:23.470306 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:23 crc kubenswrapper[4970]: I1124 13:48:23.470330 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:23 crc kubenswrapper[4970]: E1124 13:48:23.471133 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.115768 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/1.log" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.117130 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/0.log" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.117268 4970 generic.go:334] "Generic (PLEG): container finished" podID="0875a754-321a-4a3a-a798-941eefcdffe5" containerID="53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a" exitCode=1 Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.117360 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerDied","Data":"53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a"} Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.117458 4970 scope.go:117] "RemoveContainer" containerID="4ea054595f257f92f023b767861e11ce854a91931fc1266cbfee9b75c571f641" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.118085 4970 scope.go:117] "RemoveContainer" containerID="53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a" Nov 24 13:48:24 crc kubenswrapper[4970]: E1124 13:48:24.118445 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-vtfnd_openshift-multus(0875a754-321a-4a3a-a798-941eefcdffe5)\"" pod="openshift-multus/multus-vtfnd" podUID="0875a754-321a-4a3a-a798-941eefcdffe5" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.147399 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5wbs6" podStartSLOduration=95.147328059 podStartE2EDuration="1m35.147328059s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:14.104933248 +0000 UTC m=+109.392690571" watchObservedRunningTime="2025-11-24 13:48:24.147328059 +0000 UTC m=+119.435085392" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.469394 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:24 crc kubenswrapper[4970]: I1124 13:48:24.469399 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:24 crc kubenswrapper[4970]: E1124 13:48:24.469572 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:24 crc kubenswrapper[4970]: E1124 13:48:24.469721 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:25 crc kubenswrapper[4970]: I1124 13:48:25.123327 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/1.log" Nov 24 13:48:25 crc kubenswrapper[4970]: I1124 13:48:25.469815 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:25 crc kubenswrapper[4970]: I1124 13:48:25.469880 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:25 crc kubenswrapper[4970]: E1124 13:48:25.471504 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:25 crc kubenswrapper[4970]: E1124 13:48:25.471724 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:25 crc kubenswrapper[4970]: E1124 13:48:25.484103 4970 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 13:48:25 crc kubenswrapper[4970]: E1124 13:48:25.576897 4970 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:48:26 crc kubenswrapper[4970]: I1124 13:48:26.470117 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:26 crc kubenswrapper[4970]: E1124 13:48:26.470273 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:26 crc kubenswrapper[4970]: I1124 13:48:26.470909 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:26 crc kubenswrapper[4970]: E1124 13:48:26.471153 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:27 crc kubenswrapper[4970]: I1124 13:48:27.469847 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:27 crc kubenswrapper[4970]: I1124 13:48:27.469987 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:27 crc kubenswrapper[4970]: E1124 13:48:27.470041 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:27 crc kubenswrapper[4970]: E1124 13:48:27.470267 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:28 crc kubenswrapper[4970]: I1124 13:48:28.470036 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:28 crc kubenswrapper[4970]: I1124 13:48:28.470159 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:28 crc kubenswrapper[4970]: E1124 13:48:28.470266 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:28 crc kubenswrapper[4970]: E1124 13:48:28.470388 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:29 crc kubenswrapper[4970]: I1124 13:48:29.470020 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:29 crc kubenswrapper[4970]: I1124 13:48:29.470103 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:29 crc kubenswrapper[4970]: E1124 13:48:29.470246 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:29 crc kubenswrapper[4970]: E1124 13:48:29.470339 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:30 crc kubenswrapper[4970]: I1124 13:48:30.470312 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:30 crc kubenswrapper[4970]: I1124 13:48:30.470438 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:30 crc kubenswrapper[4970]: E1124 13:48:30.471083 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:30 crc kubenswrapper[4970]: E1124 13:48:30.471279 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:30 crc kubenswrapper[4970]: I1124 13:48:30.471552 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:48:30 crc kubenswrapper[4970]: E1124 13:48:30.578377 4970 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.151796 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.153964 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerStarted","Data":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.154875 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.184424 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podStartSLOduration=102.184396887 podStartE2EDuration="1m42.184396887s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:31.183985716 +0000 UTC m=+126.471743029" watchObservedRunningTime="2025-11-24 13:48:31.184396887 +0000 UTC m=+126.472154220" Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.409360 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kv7zv"] Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.409552 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:31 crc kubenswrapper[4970]: E1124 13:48:31.409782 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:31 crc kubenswrapper[4970]: I1124 13:48:31.470286 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:31 crc kubenswrapper[4970]: E1124 13:48:31.470466 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:32 crc kubenswrapper[4970]: I1124 13:48:32.469376 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:32 crc kubenswrapper[4970]: I1124 13:48:32.469454 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:32 crc kubenswrapper[4970]: E1124 13:48:32.469964 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:32 crc kubenswrapper[4970]: E1124 13:48:32.470166 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:33 crc kubenswrapper[4970]: I1124 13:48:33.470200 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:33 crc kubenswrapper[4970]: I1124 13:48:33.470373 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:33 crc kubenswrapper[4970]: E1124 13:48:33.470461 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:33 crc kubenswrapper[4970]: E1124 13:48:33.470553 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:34 crc kubenswrapper[4970]: I1124 13:48:34.469823 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:34 crc kubenswrapper[4970]: I1124 13:48:34.469840 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:34 crc kubenswrapper[4970]: E1124 13:48:34.470000 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:34 crc kubenswrapper[4970]: E1124 13:48:34.470132 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:35 crc kubenswrapper[4970]: I1124 13:48:35.469739 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:35 crc kubenswrapper[4970]: I1124 13:48:35.469784 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:35 crc kubenswrapper[4970]: E1124 13:48:35.471693 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:35 crc kubenswrapper[4970]: E1124 13:48:35.471830 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:35 crc kubenswrapper[4970]: E1124 13:48:35.579743 4970 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 13:48:36 crc kubenswrapper[4970]: I1124 13:48:36.470452 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:36 crc kubenswrapper[4970]: I1124 13:48:36.470933 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:36 crc kubenswrapper[4970]: E1124 13:48:36.470969 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:36 crc kubenswrapper[4970]: E1124 13:48:36.471455 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:37 crc kubenswrapper[4970]: I1124 13:48:37.470183 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:37 crc kubenswrapper[4970]: I1124 13:48:37.470267 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:37 crc kubenswrapper[4970]: E1124 13:48:37.470438 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:37 crc kubenswrapper[4970]: E1124 13:48:37.470637 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:38 crc kubenswrapper[4970]: I1124 13:48:38.469991 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:38 crc kubenswrapper[4970]: I1124 13:48:38.470019 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:38 crc kubenswrapper[4970]: E1124 13:48:38.470130 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:38 crc kubenswrapper[4970]: I1124 13:48:38.470654 4970 scope.go:117] "RemoveContainer" containerID="53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a" Nov 24 13:48:38 crc kubenswrapper[4970]: E1124 13:48:38.470697 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:39 crc kubenswrapper[4970]: I1124 13:48:39.191383 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/1.log" Nov 24 13:48:39 crc kubenswrapper[4970]: I1124 13:48:39.191843 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerStarted","Data":"3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde"} Nov 24 13:48:39 crc kubenswrapper[4970]: I1124 13:48:39.469729 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:39 crc kubenswrapper[4970]: I1124 13:48:39.469819 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:39 crc kubenswrapper[4970]: E1124 13:48:39.469913 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 13:48:39 crc kubenswrapper[4970]: E1124 13:48:39.470000 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kv7zv" podUID="8b0fceff-420b-4465-9afa-99af03170306" Nov 24 13:48:40 crc kubenswrapper[4970]: I1124 13:48:40.469682 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:40 crc kubenswrapper[4970]: I1124 13:48:40.469718 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:40 crc kubenswrapper[4970]: E1124 13:48:40.469914 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 13:48:40 crc kubenswrapper[4970]: E1124 13:48:40.470239 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.469522 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.469553 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.475003 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.475142 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.475021 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 13:48:41 crc kubenswrapper[4970]: I1124 13:48:41.475542 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 13:48:42 crc kubenswrapper[4970]: I1124 13:48:42.470334 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:42 crc kubenswrapper[4970]: I1124 13:48:42.470371 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:42 crc kubenswrapper[4970]: I1124 13:48:42.473911 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 13:48:42 crc kubenswrapper[4970]: I1124 13:48:42.474348 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.078221 4970 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.128854 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.129229 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.131615 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.131943 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.132339 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.133559 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.137333 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvg4n"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.138015 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.141312 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.147348 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.148384 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.155253 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.155307 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.155335 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.155459 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-d49vj"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.155852 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156181 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156289 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156427 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156666 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156784 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.156688 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157207 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157390 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157393 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157592 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.158357 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157447 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157463 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.157480 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.158248 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.159032 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5jd8l"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.166130 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.180994 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.181411 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.182270 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.182568 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.182908 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.183065 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.183075 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.183627 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.183895 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184047 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184097 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184122 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184142 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184183 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wctw9\" (UniqueName: \"kubernetes.io/projected/2d781049-65f3-4789-9253-21f620c2827b-kube-api-access-wctw9\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184219 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-config\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184293 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184309 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184406 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d781049-65f3-4789-9253-21f620c2827b-serving-cert\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184430 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ntcq\" (UniqueName: \"kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184452 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.184716 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.186420 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.186775 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.186941 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.187378 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.189260 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.189451 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.189555 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.189712 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.190216 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.190604 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.190766 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.191046 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.191810 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192058 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192094 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192250 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192292 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192063 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192453 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192494 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192502 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.192682 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.193182 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.193479 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.193852 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.193989 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.194277 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8pltx"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.194846 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.195643 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.196263 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.198306 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.198982 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.199395 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72mjc"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.200036 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.202405 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-5s9hm"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.203174 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qz272"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.203885 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.204002 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.204315 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.204443 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.207457 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.208504 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.208891 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.208942 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.209103 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.209162 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.209548 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.209914 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.209916 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.210500 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.210548 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.212304 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.212537 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.215347 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.215347 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.215989 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.216058 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.216254 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.216284 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.215995 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.216831 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.216957 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.222192 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.225606 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.226308 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.226533 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.226527 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.259795 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.259993 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5jd8l"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.260021 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-d49vj"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.261415 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.261569 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.262001 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.262397 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rd2x8"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.262601 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.263107 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.263611 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.263760 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.263896 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264019 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264146 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264253 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264303 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264478 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264618 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264727 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264837 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264963 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.265030 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.264662 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.269304 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.270183 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.270439 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.270748 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.270737 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.271084 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.273388 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lql7t"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.273890 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.273970 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.274327 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.274663 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.274893 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.275064 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.275196 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.276138 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.282633 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284823 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284863 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-audit-policies\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284904 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284932 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284959 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.284988 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285012 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-auth-proxy-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285038 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de6a04c8-bd9a-4563-85db-e83ba104deb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285067 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wctw9\" (UniqueName: \"kubernetes.io/projected/2d781049-65f3-4789-9253-21f620c2827b-kube-api-access-wctw9\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285095 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285118 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2vjw\" (UniqueName: \"kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285144 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09957632-4c30-4a0d-af06-ef1f12b9dff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285169 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-machine-approver-tls\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285191 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285215 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggls4\" (UniqueName: \"kubernetes.io/projected/289231a8-1bd6-4f94-b88f-6f4273e242bc-kube-api-access-ggls4\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285239 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6djdq\" (UniqueName: \"kubernetes.io/projected/835318f4-7393-4594-baab-55d2ab371714-kube-api-access-6djdq\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285262 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbvd\" (UniqueName: \"kubernetes.io/projected/09957632-4c30-4a0d-af06-ef1f12b9dff9-kube-api-access-cmbvd\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285291 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285314 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkw7r\" (UniqueName: \"kubernetes.io/projected/550abb24-6e1a-4f62-a506-ebb047eb6b92-kube-api-access-xkw7r\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285342 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3999597f-f349-4448-b4f4-45fa925f3342-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285381 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285409 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ntcq\" (UniqueName: \"kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285426 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285433 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-config\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285460 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit-dir\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285485 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285507 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-images\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285530 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285559 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285605 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnrm9\" (UniqueName: \"kubernetes.io/projected/cf6fcada-acb9-47fa-9ec2-99283583d712-kube-api-access-pnrm9\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285630 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285662 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-etcd-client\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285685 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285710 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-image-import-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285763 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285800 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2knwk\" (UniqueName: \"kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285824 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285851 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-trusted-ca\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285876 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285902 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285927 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/09957632-4c30-4a0d-af06-ef1f12b9dff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285952 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf6fcada-acb9-47fa-9ec2-99283583d712-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285975 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-serving-cert\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286002 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286028 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286051 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286079 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286103 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286131 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cm6q\" (UniqueName: \"kubernetes.io/projected/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-kube-api-access-5cm6q\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286171 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-node-pullsecrets\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286197 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286220 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msw6g\" (UniqueName: \"kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286253 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286280 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286302 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shvg8\" (UniqueName: \"kubernetes.io/projected/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-kube-api-access-shvg8\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286327 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8clv\" (UniqueName: \"kubernetes.io/projected/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-kube-api-access-z8clv\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286350 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-encryption-config\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286373 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286400 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-config\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286426 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286452 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-client\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286479 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286506 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286555 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvbt8\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-kube-api-access-kvbt8\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286629 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286661 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d781049-65f3-4789-9253-21f620c2827b-serving-cert\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286688 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6fcada-acb9-47fa-9ec2-99283583d712-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286711 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/de6a04c8-bd9a-4563-85db-e83ba104deb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286741 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286765 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/835318f4-7393-4594-baab-55d2ab371714-audit-dir\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286786 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-serving-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286819 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286841 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-serving-cert\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286867 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-encryption-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286890 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286917 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286942 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286970 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286994 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287020 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgkbv\" (UniqueName: \"kubernetes.io/projected/3999597f-f349-4448-b4f4-45fa925f3342-kube-api-access-wgkbv\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287043 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287067 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550abb24-6e1a-4f62-a506-ebb047eb6b92-serving-cert\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287090 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-config\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287115 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l28mm\" (UniqueName: \"kubernetes.io/projected/f583f773-9909-4b44-8c08-f4fbd14830be-kube-api-access-l28mm\") pod \"downloads-7954f5f757-5s9hm\" (UID: \"f583f773-9909-4b44-8c08-f4fbd14830be\") " pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287141 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.287168 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpmxm\" (UniqueName: \"kubernetes.io/projected/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-kube-api-access-dpmxm\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.285663 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.288152 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.288899 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.289561 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.290323 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.295916 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.296305 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.296498 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-nhtn4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.286298 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.298873 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvg4n"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.298985 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.302206 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mv65q"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.302716 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d781049-65f3-4789-9253-21f620c2827b-serving-cert\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.302761 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.303189 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.303374 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.303688 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.304042 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.304240 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-config\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.304436 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d781049-65f3-4789-9253-21f620c2827b-service-ca-bundle\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.304527 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.305115 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.309700 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.330501 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.334286 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.334337 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.334613 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.341873 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.352221 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kxz9f"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.352477 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.352973 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.353854 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.354495 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.355284 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.355874 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.356797 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.357589 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.357973 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.358477 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.359183 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.359713 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.360642 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.361438 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.361945 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.362616 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.363451 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.364924 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qz272"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.365995 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.367019 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.368028 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rd2x8"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.370542 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.370660 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.371738 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5s9hm"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.373139 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.374237 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8pltx"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.376787 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.376817 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.377522 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.378672 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.378713 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vf658"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.379431 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.380224 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nhtn4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.381378 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mv65q"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.382389 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.383421 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.384527 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.385587 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.386674 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387665 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-config\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387693 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit-dir\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387711 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387728 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnrm9\" (UniqueName: \"kubernetes.io/projected/cf6fcada-acb9-47fa-9ec2-99283583d712-kube-api-access-pnrm9\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387744 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387767 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72mjc"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387769 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-etcd-client\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387872 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387894 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387920 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2knwk\" (UniqueName: \"kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387939 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387959 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387974 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387992 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/09957632-4c30-4a0d-af06-ef1f12b9dff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388011 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf6fcada-acb9-47fa-9ec2-99283583d712-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388053 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-serving-cert\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388089 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388108 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cm6q\" (UniqueName: \"kubernetes.io/projected/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-kube-api-access-5cm6q\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388135 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-node-pullsecrets\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388155 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388183 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388187 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388199 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msw6g\" (UniqueName: \"kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.387788 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit-dir\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388217 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388237 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8clv\" (UniqueName: \"kubernetes.io/projected/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-kube-api-access-z8clv\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388254 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388275 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388292 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-metrics-certs\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388313 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-client\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388328 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388347 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388364 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ba805cc-d03e-4b6a-9ae7-1392d4253730-metrics-tls\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388386 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk56x\" (UniqueName: \"kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388401 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388416 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/835318f4-7393-4594-baab-55d2ab371714-audit-dir\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388432 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-serving-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388450 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388466 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-serving-cert\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388481 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388499 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388511 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-config\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388516 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-config\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388557 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvtrj\" (UniqueName: \"kubernetes.io/projected/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-kube-api-access-gvtrj\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388594 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388611 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fda55b3-c70d-40cd-8c4d-69d838eb733f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388632 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-audit-policies\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388647 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388665 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388681 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-service-ca-bundle\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388697 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ttks\" (UniqueName: \"kubernetes.io/projected/8d68406f-842a-4068-97e1-8efa425c71be-kube-api-access-9ttks\") pod \"migrator-59844c95c7-s259n\" (UID: \"8d68406f-842a-4068-97e1-8efa425c71be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388786 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388809 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-auth-proxy-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388828 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de6a04c8-bd9a-4563-85db-e83ba104deb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388847 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388863 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2vjw\" (UniqueName: \"kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388880 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-machine-approver-tls\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388895 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388911 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388925 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388916 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbvd\" (UniqueName: \"kubernetes.io/projected/09957632-4c30-4a0d-af06-ef1f12b9dff9-kube-api-access-cmbvd\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388963 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.388987 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4b2d5731-75aa-457d-8afb-75815fd5ca93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389023 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz86l\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-kube-api-access-mz86l\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389051 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389074 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-service-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389096 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-metrics-tls\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389119 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-srv-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389144 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389166 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-images\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389190 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-client\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389215 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-image-import-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389236 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwxrj\" (UniqueName: \"kubernetes.io/projected/5fda55b3-c70d-40cd-8c4d-69d838eb733f-kube-api-access-fwxrj\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389268 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/4b2d5731-75aa-457d-8afb-75815fd5ca93-kube-api-access-7n7v6\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389295 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-trusted-ca\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389319 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389341 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-stats-auth\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389363 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389384 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389408 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389430 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shvg8\" (UniqueName: \"kubernetes.io/projected/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-kube-api-access-shvg8\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389454 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-default-certificate\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389475 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-encryption-config\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389485 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-audit-policies\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389509 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-serving-cert\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389535 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mcqw\" (UniqueName: \"kubernetes.io/projected/9ba805cc-d03e-4b6a-9ae7-1392d4253730-kube-api-access-5mcqw\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389558 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-cert\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389604 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-profile-collector-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389633 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389658 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvbt8\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-kube-api-access-kvbt8\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389682 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6chg\" (UniqueName: \"kubernetes.io/projected/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-kube-api-access-k6chg\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389742 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6fcada-acb9-47fa-9ec2-99283583d712-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389766 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/de6a04c8-bd9a-4563-85db-e83ba104deb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389790 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389813 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389835 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrpt\" (UniqueName: \"kubernetes.io/projected/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-kube-api-access-vrrpt\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389859 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5ckl\" (UniqueName: \"kubernetes.io/projected/95d62494-ded0-4f72-b037-d8f47de79dd2-kube-api-access-d5ckl\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389893 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-encryption-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389918 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389941 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389963 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550abb24-6e1a-4f62-a506-ebb047eb6b92-serving-cert\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389986 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgkbv\" (UniqueName: \"kubernetes.io/projected/3999597f-f349-4448-b4f4-45fa925f3342-kube-api-access-wgkbv\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390007 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390031 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l28mm\" (UniqueName: \"kubernetes.io/projected/f583f773-9909-4b44-8c08-f4fbd14830be-kube-api-access-l28mm\") pod \"downloads-7954f5f757-5s9hm\" (UID: \"f583f773-9909-4b44-8c08-f4fbd14830be\") " pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390055 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390078 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpmxm\" (UniqueName: \"kubernetes.io/projected/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-kube-api-access-dpmxm\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390101 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-config\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390130 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390156 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggls4\" (UniqueName: \"kubernetes.io/projected/289231a8-1bd6-4f94-b88f-6f4273e242bc-kube-api-access-ggls4\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390179 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390202 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390227 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09957632-4c30-4a0d-af06-ef1f12b9dff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390253 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6djdq\" (UniqueName: \"kubernetes.io/projected/835318f4-7393-4594-baab-55d2ab371714-kube-api-access-6djdq\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390275 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkw7r\" (UniqueName: \"kubernetes.io/projected/550abb24-6e1a-4f62-a506-ebb047eb6b92-kube-api-access-xkw7r\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390300 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3999597f-f349-4448-b4f4-45fa925f3342-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390324 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390337 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390350 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjdgv\" (UniqueName: \"kubernetes.io/projected/5a222693-bf87-4da6-aff7-5b13c6234a73-kube-api-access-qjdgv\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390384 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-trusted-ca\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390410 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.390434 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.391295 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.391313 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de6a04c8-bd9a-4563-85db-e83ba104deb2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.391500 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.391800 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-auth-proxy-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.391966 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.392323 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.392418 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-etcd-client\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.392445 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.392470 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vf658"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.392480 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s97wf"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.393008 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/835318f4-7393-4594-baab-55d2ab371714-audit-dir\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.393630 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.393889 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/09957632-4c30-4a0d-af06-ef1f12b9dff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.394682 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf6fcada-acb9-47fa-9ec2-99283583d712-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.395222 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.395390 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.389269 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-config\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.395887 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-audit\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.396658 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.396767 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/289231a8-1bd6-4f94-b88f-6f4273e242bc-node-pullsecrets\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.396704 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-serving-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.396978 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.397217 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-etcd-client\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.397524 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.397815 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.398392 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/835318f4-7393-4594-baab-55d2ab371714-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.397873 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.398813 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-trusted-ca-bundle\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.398929 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-machine-approver-tls\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.398937 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.399025 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/289231a8-1bd6-4f94-b88f-6f4273e242bc-image-import-ca\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.399214 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/550abb24-6e1a-4f62-a506-ebb047eb6b92-trusted-ca\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.399691 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3999597f-f349-4448-b4f4-45fa925f3342-images\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.399894 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.400535 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.400634 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-config\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.400765 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.400857 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-tpqrz"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.401037 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.401619 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.402027 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.402342 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.403315 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-encryption-config\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.403350 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-serving-cert\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404099 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404373 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404381 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404422 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404496 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404913 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.404966 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.406972 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/de6a04c8-bd9a-4563-85db-e83ba104deb2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.407097 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6fcada-acb9-47fa-9ec2-99283583d712-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.407479 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09957632-4c30-4a0d-af06-ef1f12b9dff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.408328 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.409489 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550abb24-6e1a-4f62-a506-ebb047eb6b92-serving-cert\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.409960 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.410336 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/289231a8-1bd6-4f94-b88f-6f4273e242bc-encryption-config\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.410356 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/835318f4-7393-4594-baab-55d2ab371714-serving-cert\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.411762 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.412318 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.413948 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.417397 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.417900 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3999597f-f349-4448-b4f4-45fa925f3342-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.418473 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.419438 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kxz9f"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.420473 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.421409 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.422423 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s97wf"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.423372 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.424331 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rsrgq"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.425039 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.425313 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rsrgq"] Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.431049 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.450623 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.490477 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.490994 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491033 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-metrics-certs\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491052 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491073 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ba805cc-d03e-4b6a-9ae7-1392d4253730-metrics-tls\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491097 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk56x\" (UniqueName: \"kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491130 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491156 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvtrj\" (UniqueName: \"kubernetes.io/projected/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-kube-api-access-gvtrj\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491179 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491201 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fda55b3-c70d-40cd-8c4d-69d838eb733f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491226 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-service-ca-bundle\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491250 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ttks\" (UniqueName: \"kubernetes.io/projected/8d68406f-842a-4068-97e1-8efa425c71be-kube-api-access-9ttks\") pod \"migrator-59844c95c7-s259n\" (UID: \"8d68406f-842a-4068-97e1-8efa425c71be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491304 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4b2d5731-75aa-457d-8afb-75815fd5ca93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491341 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz86l\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-kube-api-access-mz86l\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491367 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-service-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491392 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-metrics-tls\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491416 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-srv-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491440 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-client\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491466 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwxrj\" (UniqueName: \"kubernetes.io/projected/5fda55b3-c70d-40cd-8c4d-69d838eb733f-kube-api-access-fwxrj\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491501 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/4b2d5731-75aa-457d-8afb-75815fd5ca93-kube-api-access-7n7v6\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491530 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-stats-auth\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491556 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491646 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-default-certificate\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491674 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-serving-cert\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491699 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mcqw\" (UniqueName: \"kubernetes.io/projected/9ba805cc-d03e-4b6a-9ae7-1392d4253730-kube-api-access-5mcqw\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491725 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-cert\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491750 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-profile-collector-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491785 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6chg\" (UniqueName: \"kubernetes.io/projected/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-kube-api-access-k6chg\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491813 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491839 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrpt\" (UniqueName: \"kubernetes.io/projected/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-kube-api-access-vrrpt\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491863 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5ckl\" (UniqueName: \"kubernetes.io/projected/95d62494-ded0-4f72-b037-d8f47de79dd2-kube-api-access-d5ckl\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491933 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-config\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491967 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491989 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492030 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.491989 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-service-ca-bundle\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492054 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjdgv\" (UniqueName: \"kubernetes.io/projected/5a222693-bf87-4da6-aff7-5b13c6234a73-kube-api-access-qjdgv\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492081 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-trusted-ca\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492108 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492134 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.492196 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.494264 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ba805cc-d03e-4b6a-9ae7-1392d4253730-metrics-tls\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.510631 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.531635 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.535140 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-default-certificate\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.551336 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.555680 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-stats-auth\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.571064 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.573875 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-metrics-certs\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.593501 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.611375 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.630918 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.651192 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.671140 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.709098 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wctw9\" (UniqueName: \"kubernetes.io/projected/2d781049-65f3-4789-9253-21f620c2827b-kube-api-access-wctw9\") pod \"authentication-operator-69f744f599-zvg4n\" (UID: \"2d781049-65f3-4789-9253-21f620c2827b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.728813 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ntcq\" (UniqueName: \"kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq\") pod \"controller-manager-879f6c89f-rnm9t\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.731351 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.758406 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.761098 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.765431 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-trusted-ca\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.771407 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.781643 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-metrics-tls\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.782822 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.792476 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.811666 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.832396 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.852661 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.900288 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.901792 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.907698 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-cert\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.911358 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.931476 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.951447 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.970689 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.991706 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 13:48:44 crc kubenswrapper[4970]: I1124 13:48:44.996443 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-zvg4n"] Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.011682 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.012387 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: W1124 13:48:45.019137 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7f2f7bc_97c0_4e99_8a46_32815fa43b5b.slice/crio-201ced5d4f98e6653ae6fd0a1b92a36d6b365612a2c93a33cdcef205514dcc5d WatchSource:0}: Error finding container 201ced5d4f98e6653ae6fd0a1b92a36d6b365612a2c93a33cdcef205514dcc5d: Status 404 returned error can't find the container with id 201ced5d4f98e6653ae6fd0a1b92a36d6b365612a2c93a33cdcef205514dcc5d Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.025348 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-serving-cert\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.030959 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.034521 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.051313 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.052177 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-service-ca\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.071321 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.075319 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-etcd-client\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.091905 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.111286 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.112942 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-config\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.130263 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.136826 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-profile-collector-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.150705 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.171266 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.191277 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.211773 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.212523 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" event={"ID":"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b","Type":"ContainerStarted","Data":"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c"} Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.212633 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" event={"ID":"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b","Type":"ContainerStarted","Data":"201ced5d4f98e6653ae6fd0a1b92a36d6b365612a2c93a33cdcef205514dcc5d"} Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.212981 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.214758 4970 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-rnm9t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.214836 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.224753 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" event={"ID":"2d781049-65f3-4789-9253-21f620c2827b","Type":"ContainerStarted","Data":"53736393658f00651f69793bbc2b4a0f09ddf74e1245b7a27825366d02d052a8"} Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.224824 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" event={"ID":"2d781049-65f3-4789-9253-21f620c2827b","Type":"ContainerStarted","Data":"d34768d9a404b5221341e3713bd532d8d820b97b5c274950c4ceb1effbeec2c8"} Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.231527 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.250993 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.256531 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5a222693-bf87-4da6-aff7-5b13c6234a73-srv-cert\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.270774 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.291077 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.312258 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.330326 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.349598 4970 request.go:700] Waited for 1.006477128s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-tls&limit=500&resourceVersion=0 Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.351507 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.356664 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4b2d5731-75aa-457d-8afb-75815fd5ca93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.370891 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.391426 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.411719 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.431202 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.450991 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.471829 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.488240 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5fda55b3-c70d-40cd-8c4d-69d838eb733f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491217 4970 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491312 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert podName:de4a90c4-c8e5-45f2-b8a9-419acf2d26a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.991287645 +0000 UTC m=+141.279044958 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert") pod "kube-apiserver-operator-766d6c64bb-6cnrs" (UID: "de4a90c4-c8e5-45f2-b8a9-419acf2d26a5") : failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491385 4970 secret.go:188] Couldn't get secret openshift-kube-scheduler-operator/kube-scheduler-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491416 4970 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491504 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert podName:8f376ff7-d945-45e1-b6ff-df9769b1ecb4 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.991470909 +0000 UTC m=+141.279228242 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert") pod "openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" (UID: "8f376ff7-d945-45e1-b6ff-df9769b1ecb4") : failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491537 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert podName:95d62494-ded0-4f72-b037-d8f47de79dd2 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.991522661 +0000 UTC m=+141.279279984 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert") pod "service-ca-operator-777779d784-t9qzb" (UID: "95d62494-ded0-4f72-b037-d8f47de79dd2") : failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491814 4970 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491866 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config podName:95d62494-ded0-4f72-b037-d8f47de79dd2 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.991854769 +0000 UTC m=+141.279612072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config") pod "service-ca-operator-777779d784-t9qzb" (UID: "95d62494-ded0-4f72-b037-d8f47de79dd2") : failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.491941 4970 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492005 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca podName:8b04d05d-1031-4197-8fb3-59c52252bad7 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.991991202 +0000 UTC m=+141.279748535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca") pod "marketplace-operator-79b997595-m968g" (UID: "8b04d05d-1031-4197-8fb3-59c52252bad7") : failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.492060 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492640 4970 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492639 4970 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492683 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config podName:de4a90c4-c8e5-45f2-b8a9-419acf2d26a5 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.99267164 +0000 UTC m=+141.280428943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config") pod "kube-apiserver-operator-766d6c64bb-6cnrs" (UID: "de4a90c4-c8e5-45f2-b8a9-419acf2d26a5") : failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492740 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics podName:8b04d05d-1031-4197-8fb3-59c52252bad7 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.992707411 +0000 UTC m=+141.280464774 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics") pod "marketplace-operator-79b997595-m968g" (UID: "8b04d05d-1031-4197-8fb3-59c52252bad7") : failed to sync secret cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492753 4970 configmap.go:193] Couldn't get configMap openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-config: failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: E1124 13:48:45.492862 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config podName:8f376ff7-d945-45e1-b6ff-df9769b1ecb4 nodeName:}" failed. No retries permitted until 2025-11-24 13:48:45.992828024 +0000 UTC m=+141.280585387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config") pod "openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" (UID: "8f376ff7-d945-45e1-b6ff-df9769b1ecb4") : failed to sync configmap cache: timed out waiting for the condition Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.511499 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.530817 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.550995 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.570598 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.591901 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.620023 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.632046 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.651259 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.671411 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.691279 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.713888 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.731618 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.750976 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.772424 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.790900 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.811027 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.831135 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.851015 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.892702 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.911663 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.931598 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.950960 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.971691 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 13:48:45 crc kubenswrapper[4970]: I1124 13:48:45.990931 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.031239 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.034903 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035016 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035184 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035231 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035338 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035392 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035464 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.035529 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.041016 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95d62494-ded0-4f72-b037-d8f47de79dd2-serving-cert\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.042067 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95d62494-ded0-4f72-b037-d8f47de79dd2-config\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.042179 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-config\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.043396 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.043553 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.045413 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.045461 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.047021 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.055371 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.073325 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbvd\" (UniqueName: \"kubernetes.io/projected/09957632-4c30-4a0d-af06-ef1f12b9dff9-kube-api-access-cmbvd\") pod \"openshift-config-operator-7777fb866f-72mjc\" (UID: \"09957632-4c30-4a0d-af06-ef1f12b9dff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.084730 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnrm9\" (UniqueName: \"kubernetes.io/projected/cf6fcada-acb9-47fa-9ec2-99283583d712-kube-api-access-pnrm9\") pod \"openshift-controller-manager-operator-756b6f6bc6-hbt4q\" (UID: \"cf6fcada-acb9-47fa-9ec2-99283583d712\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.118533 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2knwk\" (UniqueName: \"kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk\") pod \"route-controller-manager-6576b87f9c-c925s\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.127217 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msw6g\" (UniqueName: \"kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g\") pod \"console-f9d7485db-fsgxr\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.147076 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8clv\" (UniqueName: \"kubernetes.io/projected/2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7-kube-api-access-z8clv\") pod \"openshift-apiserver-operator-796bbdcf4f-tfg84\" (UID: \"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.147859 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.154613 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.176041 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2vjw\" (UniqueName: \"kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw\") pod \"oauth-openshift-558db77b4-lr6pb\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.177065 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.192800 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.196079 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cm6q\" (UniqueName: \"kubernetes.io/projected/0335a407-86a1-4ff0-9e46-df9f7ce8f3cc-kube-api-access-5cm6q\") pod \"multus-admission-controller-857f4d67dd-qz272\" (UID: \"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.226386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shvg8\" (UniqueName: \"kubernetes.io/projected/decc8d43-ce9c-4d67-a2a5-c4203bbc12bb-kube-api-access-shvg8\") pod \"machine-approver-56656f9798-pw2s5\" (UID: \"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.235397 4970 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.236716 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.255562 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.292673 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l28mm\" (UniqueName: \"kubernetes.io/projected/f583f773-9909-4b44-8c08-f4fbd14830be-kube-api-access-l28mm\") pod \"downloads-7954f5f757-5s9hm\" (UID: \"f583f773-9909-4b44-8c08-f4fbd14830be\") " pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.318539 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.319140 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgkbv\" (UniqueName: \"kubernetes.io/projected/3999597f-f349-4448-b4f4-45fa925f3342-kube-api-access-wgkbv\") pod \"machine-api-operator-5694c8668f-5jd8l\" (UID: \"3999597f-f349-4448-b4f4-45fa925f3342\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.323874 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.342021 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggls4\" (UniqueName: \"kubernetes.io/projected/289231a8-1bd6-4f94-b88f-6f4273e242bc-kube-api-access-ggls4\") pod \"apiserver-76f77b778f-d49vj\" (UID: \"289231a8-1bd6-4f94-b88f-6f4273e242bc\") " pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.356503 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpmxm\" (UniqueName: \"kubernetes.io/projected/cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f-kube-api-access-dpmxm\") pod \"cluster-samples-operator-665b6dd947-sxctt\" (UID: \"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.362622 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.370664 4970 request.go:700] Waited for 1.969913379s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.373661 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.382255 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6djdq\" (UniqueName: \"kubernetes.io/projected/835318f4-7393-4594-baab-55d2ab371714-kube-api-access-6djdq\") pod \"apiserver-7bbb656c7d-hz58r\" (UID: \"835318f4-7393-4594-baab-55d2ab371714\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.384716 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.409351 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.419269 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkw7r\" (UniqueName: \"kubernetes.io/projected/550abb24-6e1a-4f62-a506-ebb047eb6b92-kube-api-access-xkw7r\") pod \"console-operator-58897d9998-8pltx\" (UID: \"550abb24-6e1a-4f62-a506-ebb047eb6b92\") " pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.432384 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvbt8\" (UniqueName: \"kubernetes.io/projected/de6a04c8-bd9a-4563-85db-e83ba104deb2-kube-api-access-kvbt8\") pod \"cluster-image-registry-operator-dc59b4c8b-p579k\" (UID: \"de6a04c8-bd9a-4563-85db-e83ba104deb2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.434725 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.445433 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.452494 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.460744 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.461094 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.468288 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.470835 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.489088 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.511250 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-72mjc"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.511566 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.517900 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.565912 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.591984 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.617880 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk56x\" (UniqueName: \"kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x\") pod \"marketplace-operator-79b997595-m968g\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.618841 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.622713 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvtrj\" (UniqueName: \"kubernetes.io/projected/a9761a4c-d28b-475d-a6d8-8f78c29c42f6-kube-api-access-gvtrj\") pod \"router-default-5444994796-lql7t\" (UID: \"a9761a4c-d28b-475d-a6d8-8f78c29c42f6\") " pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.655846 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.661712 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f376ff7-d945-45e1-b6ff-df9769b1ecb4-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-v8mhc\" (UID: \"8f376ff7-d945-45e1-b6ff-df9769b1ecb4\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.667418 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.678204 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ttks\" (UniqueName: \"kubernetes.io/projected/8d68406f-842a-4068-97e1-8efa425c71be-kube-api-access-9ttks\") pod \"migrator-59844c95c7-s259n\" (UID: \"8d68406f-842a-4068-97e1-8efa425c71be\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.687243 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz86l\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-kube-api-access-mz86l\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.692953 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.696485 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwxrj\" (UniqueName: \"kubernetes.io/projected/5fda55b3-c70d-40cd-8c4d-69d838eb733f-kube-api-access-fwxrj\") pod \"package-server-manager-789f6589d5-d9kzs\" (UID: \"5fda55b3-c70d-40cd-8c4d-69d838eb733f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.699759 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.709066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n7v6\" (UniqueName: \"kubernetes.io/projected/4b2d5731-75aa-457d-8afb-75815fd5ca93-kube-api-access-7n7v6\") pod \"control-plane-machine-set-operator-78cbb6b69f-xl78j\" (UID: \"4b2d5731-75aa-457d-8afb-75815fd5ca93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.732618 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mcqw\" (UniqueName: \"kubernetes.io/projected/9ba805cc-d03e-4b6a-9ae7-1392d4253730-kube-api-access-5mcqw\") pod \"dns-operator-744455d44c-rd2x8\" (UID: \"9ba805cc-d03e-4b6a-9ae7-1392d4253730\") " pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.756065 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrpt\" (UniqueName: \"kubernetes.io/projected/97b2415c-4b75-495f-bf2d-8ab1a5a23cec-kube-api-access-vrrpt\") pod \"etcd-operator-b45778765-mv65q\" (UID: \"97b2415c-4b75-495f-bf2d-8ab1a5a23cec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.777077 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6chg\" (UniqueName: \"kubernetes.io/projected/d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0-kube-api-access-k6chg\") pod \"ingress-canary-nhtn4\" (UID: \"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0\") " pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.779069 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.790620 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5ckl\" (UniqueName: \"kubernetes.io/projected/95d62494-ded0-4f72-b037-d8f47de79dd2-kube-api-access-d5ckl\") pod \"service-ca-operator-777779d784-t9qzb\" (UID: \"95d62494-ded0-4f72-b037-d8f47de79dd2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: W1124 13:48:46.791742 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod861c2ae4_caff_41f9_a437_f7702a7a5c25.slice/crio-5d094152c29dc4d8a45da75324864492547dd0f30ccdc79e17dc791714f9a0d8 WatchSource:0}: Error finding container 5d094152c29dc4d8a45da75324864492547dd0f30ccdc79e17dc791714f9a0d8: Status 404 returned error can't find the container with id 5d094152c29dc4d8a45da75324864492547dd0f30ccdc79e17dc791714f9a0d8 Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.819495 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c0a47ee9-0bab-4235-9ca4-4088ffc0241a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j7dr6\" (UID: \"c0a47ee9-0bab-4235-9ca4-4088ffc0241a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.849602 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/de4a90c4-c8e5-45f2-b8a9-419acf2d26a5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cnrs\" (UID: \"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.860926 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.864598 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.865101 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.875853 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.887479 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjdgv\" (UniqueName: \"kubernetes.io/projected/5a222693-bf87-4da6-aff7-5b13c6234a73-kube-api-access-qjdgv\") pod \"catalog-operator-68c6474976-vsj9t\" (UID: \"5a222693-bf87-4da6-aff7-5b13c6234a73\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.888192 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-nhtn4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.889623 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.889724 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjtf\" (UniqueName: \"kubernetes.io/projected/d38364f3-e04f-4789-81c8-23ea7e98d39e-kube-api-access-dnjtf\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.889842 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.889926 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f360574b-9321-4007-a976-36fb4ac6d812-config\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.889993 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-cabundle\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890131 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890211 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890284 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f360574b-9321-4007-a976-36fb4ac6d812-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890364 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890437 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d38364f3-e04f-4789-81c8-23ea7e98d39e-proxy-tls\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890517 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-apiservice-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.890601 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: E1124 13:48:46.892629 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.392609996 +0000 UTC m=+142.680367379 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900270 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktg4h\" (UniqueName: \"kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900442 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cdg8\" (UniqueName: \"kubernetes.io/projected/0819c483-76c3-48b7-9504-b8bdaeced5a5-kube-api-access-4cdg8\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900555 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-key\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900651 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0819c483-76c3-48b7-9504-b8bdaeced5a5-tmpfs\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900756 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-images\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900838 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900914 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f360574b-9321-4007-a976-36fb4ac6d812-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.900995 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.901106 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-webhook-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.901189 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7fmb\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.901258 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.901335 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.901413 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjqcf\" (UniqueName: \"kubernetes.io/projected/851664d5-b007-44c4-b4f4-99bcc0dd6054-kube-api-access-mjqcf\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.895401 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.909065 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5jd8l"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.909269 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.928800 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.929094 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.935432 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.947769 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.975278 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84"] Nov 24 13:48:46 crc kubenswrapper[4970]: I1124 13:48:46.975976 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007177 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007326 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfmwl\" (UniqueName: \"kubernetes.io/projected/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-kube-api-access-sfmwl\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007363 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cdg8\" (UniqueName: \"kubernetes.io/projected/0819c483-76c3-48b7-9504-b8bdaeced5a5-kube-api-access-4cdg8\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007400 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-key\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007417 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxzl\" (UniqueName: \"kubernetes.io/projected/a71e8ef4-11e9-4d33-9a39-256022c7c008-kube-api-access-5zxzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007443 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0819c483-76c3-48b7-9504-b8bdaeced5a5-tmpfs\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007470 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-socket-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007505 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-mountpoint-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007519 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71e8ef4-11e9-4d33-9a39-256022c7c008-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007535 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-images\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007561 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007593 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f360574b-9321-4007-a976-36fb4ac6d812-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007610 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71e8ef4-11e9-4d33-9a39-256022c7c008-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.007640 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-plugins-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.007910 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.507887187 +0000 UTC m=+142.795644520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012700 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665b8213-0520-4614-90fa-4fc8378ce70c-config-volume\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012798 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012840 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpbqw\" (UniqueName: \"kubernetes.io/projected/110f91cd-900b-4698-8b85-0dfa8d00cf57-kube-api-access-mpbqw\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012872 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-proxy-tls\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012894 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-webhook-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012967 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7fmb\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.012996 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8m4j\" (UniqueName: \"kubernetes.io/projected/665b8213-0520-4614-90fa-4fc8378ce70c-kube-api-access-f8m4j\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013017 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013048 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013071 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjqcf\" (UniqueName: \"kubernetes.io/projected/851664d5-b007-44c4-b4f4-99bcc0dd6054-kube-api-access-mjqcf\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013120 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013144 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-srv-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013187 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-certs\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013262 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013282 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/665b8213-0520-4614-90fa-4fc8378ce70c-metrics-tls\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013329 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjtf\" (UniqueName: \"kubernetes.io/projected/d38364f3-e04f-4789-81c8-23ea7e98d39e-kube-api-access-dnjtf\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013388 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013411 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f360574b-9321-4007-a976-36fb4ac6d812-config\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013427 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-node-bootstrap-token\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013450 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-cabundle\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013481 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlnx5\" (UniqueName: \"kubernetes.io/projected/6de7023f-9f6a-471b-9f6f-650e9052b9e0-kube-api-access-dlnx5\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013498 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013513 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mksv\" (UniqueName: \"kubernetes.io/projected/8816f5d7-5926-4600-8b05-0a0a6782ce85-kube-api-access-5mksv\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013547 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013564 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f360574b-9321-4007-a976-36fb4ac6d812-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013622 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013639 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d38364f3-e04f-4789-81c8-23ea7e98d39e-proxy-tls\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013673 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-registration-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013691 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-csi-data-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013753 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-apiservice-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013785 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013816 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktg4h\" (UniqueName: \"kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.013891 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.014253 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-key\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.014795 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0819c483-76c3-48b7-9504-b8bdaeced5a5-tmpfs\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.015032 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-images\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.016631 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.019466 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.022300 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.022986 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f360574b-9321-4007-a976-36fb4ac6d812-config\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.023150 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/851664d5-b007-44c4-b4f4-99bcc0dd6054-signing-cabundle\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.024390 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.025283 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.027473 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f360574b-9321-4007-a976-36fb4ac6d812-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.027542 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-webhook-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.028615 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.528598239 +0000 UTC m=+142.816355532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.033300 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.035085 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d38364f3-e04f-4789-81c8-23ea7e98d39e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.035187 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-5s9hm"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.037425 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.040254 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d38364f3-e04f-4789-81c8-23ea7e98d39e-proxy-tls\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.041236 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0819c483-76c3-48b7-9504-b8bdaeced5a5-apiservice-cert\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.043868 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.049623 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cdg8\" (UniqueName: \"kubernetes.io/projected/0819c483-76c3-48b7-9504-b8bdaeced5a5-kube-api-access-4cdg8\") pod \"packageserver-d55dfcdfc-7x2m4\" (UID: \"0819c483-76c3-48b7-9504-b8bdaeced5a5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.054652 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qz272"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.088245 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f360574b-9321-4007-a976-36fb4ac6d812-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-cxqx2\" (UID: \"f360574b-9321-4007-a976-36fb4ac6d812\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.088697 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjtf\" (UniqueName: \"kubernetes.io/projected/d38364f3-e04f-4789-81c8-23ea7e98d39e-kube-api-access-dnjtf\") pod \"machine-config-operator-74547568cd-ps7l4\" (UID: \"d38364f3-e04f-4789-81c8-23ea7e98d39e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.108370 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7fmb\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114441 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114691 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71e8ef4-11e9-4d33-9a39-256022c7c008-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114723 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71e8ef4-11e9-4d33-9a39-256022c7c008-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114750 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-plugins-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114770 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665b8213-0520-4614-90fa-4fc8378ce70c-config-volume\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114799 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpbqw\" (UniqueName: \"kubernetes.io/projected/110f91cd-900b-4698-8b85-0dfa8d00cf57-kube-api-access-mpbqw\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114825 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-proxy-tls\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114878 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8m4j\" (UniqueName: \"kubernetes.io/projected/665b8213-0520-4614-90fa-4fc8378ce70c-kube-api-access-f8m4j\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114930 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114956 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-srv-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.114976 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-certs\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115001 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/665b8213-0520-4614-90fa-4fc8378ce70c-metrics-tls\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115032 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-node-bootstrap-token\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115058 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlnx5\" (UniqueName: \"kubernetes.io/projected/6de7023f-9f6a-471b-9f6f-650e9052b9e0-kube-api-access-dlnx5\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115079 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mksv\" (UniqueName: \"kubernetes.io/projected/8816f5d7-5926-4600-8b05-0a0a6782ce85-kube-api-access-5mksv\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115110 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-registration-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115129 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-csi-data-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115185 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115211 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfmwl\" (UniqueName: \"kubernetes.io/projected/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-kube-api-access-sfmwl\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115233 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxzl\" (UniqueName: \"kubernetes.io/projected/a71e8ef4-11e9-4d33-9a39-256022c7c008-kube-api-access-5zxzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115255 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-socket-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115277 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-mountpoint-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.115369 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-mountpoint-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.115458 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.61543852 +0000 UTC m=+142.903195823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.116361 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71e8ef4-11e9-4d33-9a39-256022c7c008-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.117159 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-csi-data-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.117491 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-registration-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.117806 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.119193 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/665b8213-0520-4614-90fa-4fc8378ce70c-config-volume\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.119268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-plugins-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.119327 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/6de7023f-9f6a-471b-9f6f-650e9052b9e0-socket-dir\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.121105 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-srv-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: W1124 13:48:47.124035 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0335a407_86a1_4ff0_9e46_df9f7ce8f3cc.slice/crio-08e443e18acfdf56b8a6600d54780b8ebe3b17a886b607024fe2e1275b2d4338 WatchSource:0}: Error finding container 08e443e18acfdf56b8a6600d54780b8ebe3b17a886b607024fe2e1275b2d4338: Status 404 returned error can't find the container with id 08e443e18acfdf56b8a6600d54780b8ebe3b17a886b607024fe2e1275b2d4338 Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.124825 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8816f5d7-5926-4600-8b05-0a0a6782ce85-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.125321 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71e8ef4-11e9-4d33-9a39-256022c7c008-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.125871 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-proxy-tls\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.132141 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-certs\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.135379 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.136804 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/665b8213-0520-4614-90fa-4fc8378ce70c-metrics-tls\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.138295 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/110f91cd-900b-4698-8b85-0dfa8d00cf57-node-bootstrap-token\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.148116 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.159301 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktg4h\" (UniqueName: \"kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h\") pod \"collect-profiles-29399865-tr7d4\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.172398 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjqcf\" (UniqueName: \"kubernetes.io/projected/851664d5-b007-44c4-b4f4-99bcc0dd6054-kube-api-access-mjqcf\") pod \"service-ca-9c57cc56f-kxz9f\" (UID: \"851664d5-b007-44c4-b4f4-99bcc0dd6054\") " pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.183468 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.219122 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.219501 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.719487283 +0000 UTC m=+143.007244576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.220135 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.230479 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.253749 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mksv\" (UniqueName: \"kubernetes.io/projected/8816f5d7-5926-4600-8b05-0a0a6782ce85-kube-api-access-5mksv\") pod \"olm-operator-6b444d44fb-7xpcq\" (UID: \"8816f5d7-5926-4600-8b05-0a0a6782ce85\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.254280 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.273507 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfmwl\" (UniqueName: \"kubernetes.io/projected/611aaa5a-fc7a-43d3-a4b4-5ac2928cde95-kube-api-access-sfmwl\") pod \"machine-config-controller-84d6567774-vf658\" (UID: \"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.277349 4970 generic.go:334] "Generic (PLEG): container finished" podID="09957632-4c30-4a0d-af06-ef1f12b9dff9" containerID="e4212aff6c8cc932c60ad386e0803d56217b72c2b423699f0de4e63786b2422a" exitCode=0 Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.277955 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" event={"ID":"09957632-4c30-4a0d-af06-ef1f12b9dff9","Type":"ContainerDied","Data":"e4212aff6c8cc932c60ad386e0803d56217b72c2b423699f0de4e63786b2422a"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.278045 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" event={"ID":"09957632-4c30-4a0d-af06-ef1f12b9dff9","Type":"ContainerStarted","Data":"25c7e572c6988ab2be116f8ab386167a1832e1b0db90bd0ab9fc3d3963a8c998"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.279771 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" event={"ID":"861c2ae4-caff-41f9-a437-f7702a7a5c25","Type":"ContainerStarted","Data":"5d094152c29dc4d8a45da75324864492547dd0f30ccdc79e17dc791714f9a0d8"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.281440 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.284520 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" event={"ID":"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc","Type":"ContainerStarted","Data":"08e443e18acfdf56b8a6600d54780b8ebe3b17a886b607024fe2e1275b2d4338"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.284825 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlnx5\" (UniqueName: \"kubernetes.io/projected/6de7023f-9f6a-471b-9f6f-650e9052b9e0-kube-api-access-dlnx5\") pod \"csi-hostpathplugin-s97wf\" (UID: \"6de7023f-9f6a-471b-9f6f-650e9052b9e0\") " pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.287059 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" event={"ID":"60d6d622-9caf-4124-9226-1aed670ffb03","Type":"ContainerStarted","Data":"0d64c2a055e7f7bf5940ad934145d81b5ca51b1d41fb62999bbdfc83c3e9d531"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.287920 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8m4j\" (UniqueName: \"kubernetes.io/projected/665b8213-0520-4614-90fa-4fc8378ce70c-kube-api-access-f8m4j\") pod \"dns-default-rsrgq\" (UID: \"665b8213-0520-4614-90fa-4fc8378ce70c\") " pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.292996 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" event={"ID":"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb","Type":"ContainerStarted","Data":"8d7f60e0c161449b460a4d5391c27f3b26402ef8bc4bfbebe03f516a621c3673"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.294491 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" event={"ID":"3999597f-f349-4448-b4f4-45fa925f3342","Type":"ContainerStarted","Data":"f061a40af3cb10bc0fe4012dafe953b775f1104aef052ee70cafed6844899a21"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.296386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpbqw\" (UniqueName: \"kubernetes.io/projected/110f91cd-900b-4698-8b85-0dfa8d00cf57-kube-api-access-mpbqw\") pod \"machine-config-server-tpqrz\" (UID: \"110f91cd-900b-4698-8b85-0dfa8d00cf57\") " pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.302405 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.319263 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" event={"ID":"cf6fcada-acb9-47fa-9ec2-99283583d712","Type":"ContainerStarted","Data":"3687005e92c53344581440c88c6ca6ee33cfbd1dfd945ccc3ff2414d7b570473"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.319334 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" event={"ID":"cf6fcada-acb9-47fa-9ec2-99283583d712","Type":"ContainerStarted","Data":"d3c35d8a6469b0f0343c23154ee470e2f7a3ee9cfeaf798bd9abd98f5181d336"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.319840 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxzl\" (UniqueName: \"kubernetes.io/projected/a71e8ef4-11e9-4d33-9a39-256022c7c008-kube-api-access-5zxzl\") pod \"kube-storage-version-migrator-operator-b67b599dd-jhzg2\" (UID: \"a71e8ef4-11e9-4d33-9a39-256022c7c008\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.319853 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.319925 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.819906713 +0000 UTC m=+143.107664056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.320469 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.321043 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.821028132 +0000 UTC m=+143.108785415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.323146 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.328823 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.329035 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" event={"ID":"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f","Type":"ContainerStarted","Data":"4885c2db0a830b1147facda8747d649be31ff06eb648c4639efcff3bafe6b894"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.331239 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tpqrz" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.339183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5s9hm" event={"ID":"f583f773-9909-4b44-8c08-f4fbd14830be","Type":"ContainerStarted","Data":"8ff159576172e2d446c1afa8407abf174d07c3d23268c10458582fd62826ac6f"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.340548 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.342135 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fsgxr" event={"ID":"45d80403-6fe8-4874-a2d9-04a3838b65f1","Type":"ContainerStarted","Data":"f59069565cf50d0421f0e89c30c44525efc64706625f8028bc457696b2f623fd"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.357228 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" event={"ID":"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7","Type":"ContainerStarted","Data":"842e64b8f3a90d43f70ac869a7a4e6171ada1c64de78119c4f5f146583ab23bb"} Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.365010 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.415487 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.421384 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.422373 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:47.922358725 +0000 UTC m=+143.210116018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.438817 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-d49vj"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.522569 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.522832 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.022821776 +0000 UTC m=+143.310579069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: W1124 13:48:47.577922 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod289231a8_1bd6_4f94_b88f_6f4273e242bc.slice/crio-d070c9738f031dec901caffda2cd048ee13d21c5569eb7fbb304f9f570be3150 WatchSource:0}: Error finding container d070c9738f031dec901caffda2cd048ee13d21c5569eb7fbb304f9f570be3150: Status 404 returned error can't find the container with id d070c9738f031dec901caffda2cd048ee13d21c5569eb7fbb304f9f570be3150 Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.589584 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.624631 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.624783 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.124761596 +0000 UTC m=+143.412518889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.624825 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.625303 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.125296279 +0000 UTC m=+143.413053572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.692617 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8pltx"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.709471 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.739380 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.740311 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.240284413 +0000 UTC m=+143.528041706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.843236 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.843525 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.343514325 +0000 UTC m=+143.631271618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.922239 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hbt4q" podStartSLOduration=118.922223187 podStartE2EDuration="1m58.922223187s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:47.920461522 +0000 UTC m=+143.208218825" watchObservedRunningTime="2025-11-24 13:48:47.922223187 +0000 UTC m=+143.209980480" Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.949254 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:47 crc kubenswrapper[4970]: E1124 13:48:47.950145 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.450118034 +0000 UTC m=+143.737875327 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.953613 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc"] Nov 24 13:48:47 crc kubenswrapper[4970]: I1124 13:48:47.963043 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rd2x8"] Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.052042 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.053033 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.553010668 +0000 UTC m=+143.840768031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.059112 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" podStartSLOduration=119.059094194 podStartE2EDuration="1m59.059094194s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:48.058027116 +0000 UTC m=+143.345784409" watchObservedRunningTime="2025-11-24 13:48:48.059094194 +0000 UTC m=+143.346851487" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.153923 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.154373 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.654359131 +0000 UTC m=+143.942116424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.238304 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-zvg4n" podStartSLOduration=119.238284217 podStartE2EDuration="1m59.238284217s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:48.202535278 +0000 UTC m=+143.490292591" watchObservedRunningTime="2025-11-24 13:48:48.238284217 +0000 UTC m=+143.526041510" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.256602 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.256911 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.756900125 +0000 UTC m=+144.044657418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: W1124 13:48:48.262913 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ba805cc_d03e_4b6a_9ae7_1392d4253730.slice/crio-c28638e11d7fc3a819a8dbb42cb7559d084822b051d145a997e50aa10d7ebfc7 WatchSource:0}: Error finding container c28638e11d7fc3a819a8dbb42cb7559d084822b051d145a997e50aa10d7ebfc7: Status 404 returned error can't find the container with id c28638e11d7fc3a819a8dbb42cb7559d084822b051d145a997e50aa10d7ebfc7 Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.359046 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.359182 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.859164802 +0000 UTC m=+144.146922095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.359271 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.359665 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.859656916 +0000 UTC m=+144.147414209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.399837 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" event={"ID":"60d6d622-9caf-4124-9226-1aed670ffb03","Type":"ContainerStarted","Data":"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.400019 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.405209 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8pltx" event={"ID":"550abb24-6e1a-4f62-a506-ebb047eb6b92","Type":"ContainerStarted","Data":"6818fac52d42eea1d0ec57a18c4a07797f8c1566ab002cce015da7cb10191276"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.414615 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tpqrz" event={"ID":"110f91cd-900b-4698-8b85-0dfa8d00cf57","Type":"ContainerStarted","Data":"b786f99be96b5f331c5b0586824118a1844d53d8a56a7a4941827dd6ecffb50e"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.418264 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" event={"ID":"8b04d05d-1031-4197-8fb3-59c52252bad7","Type":"ContainerStarted","Data":"36e439283937d8a7377b09d5a13485bc6750d1664d7fc4b76b3eb1b2322850f5"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.423402 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lql7t" event={"ID":"a9761a4c-d28b-475d-a6d8-8f78c29c42f6","Type":"ContainerStarted","Data":"afabc01af5a4d2e20faccdc7a49856239c957cd9bdec55114069607eba9bbbe3"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.425346 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" event={"ID":"835318f4-7393-4594-baab-55d2ab371714","Type":"ContainerStarted","Data":"6c1ee3470dc8f47a4315167acd530540a835221d74656f13bd946209f973b0a1"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.427523 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fsgxr" event={"ID":"45d80403-6fe8-4874-a2d9-04a3838b65f1","Type":"ContainerStarted","Data":"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.434799 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" event={"ID":"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc","Type":"ContainerStarted","Data":"4677fba6e76d56548aa4a7b59f7f487269e4e8166138896e895f85272b920edc"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.436005 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" event={"ID":"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb","Type":"ContainerStarted","Data":"92525c80c87bef60012f1c386dcf4640ad5fb23ddf151d7443e19dc7d2e6d4e9"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.437860 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" event={"ID":"09957632-4c30-4a0d-af06-ef1f12b9dff9","Type":"ContainerStarted","Data":"eae50fbe79ace1743ab1dec8b577a07365c9c1d1868104eb89ffa51706e0ce4b"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.438064 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.442121 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" event={"ID":"8f376ff7-d945-45e1-b6ff-df9769b1ecb4","Type":"ContainerStarted","Data":"ebb84bcc0b2aea89677570939841f36b29dc364995808e9e8848af80fb821f7f"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.444092 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" event={"ID":"861c2ae4-caff-41f9-a437-f7702a7a5c25","Type":"ContainerStarted","Data":"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.445672 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.448868 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" event={"ID":"9ba805cc-d03e-4b6a-9ae7-1392d4253730","Type":"ContainerStarted","Data":"c28638e11d7fc3a819a8dbb42cb7559d084822b051d145a997e50aa10d7ebfc7"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.460454 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.460626 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.960566088 +0000 UTC m=+144.248323381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.460810 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.461116 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:48.961107632 +0000 UTC m=+144.248864925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.461559 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" event={"ID":"c0a47ee9-0bab-4235-9ca4-4088ffc0241a","Type":"ContainerStarted","Data":"cd0e4f738d949d75b3e5acf4fce6e8c50a745703bdfd11b4b40d5e76e224edff"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.463378 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" event={"ID":"289231a8-1bd6-4f94-b88f-6f4273e242bc","Type":"ContainerStarted","Data":"d070c9738f031dec901caffda2cd048ee13d21c5569eb7fbb304f9f570be3150"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.466310 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" event={"ID":"2298fbfc-b73f-4b8f-bab0-1e650fa3cdf7","Type":"ContainerStarted","Data":"6691a05a6f9cf8a572433efd3812a5d8cd3adcd0a7b894c45276f023da2315d8"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.471613 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" event={"ID":"de6a04c8-bd9a-4563-85db-e83ba104deb2","Type":"ContainerStarted","Data":"259726f8d84490e9e364f40462727a1c241482e73dbf2e6a64ceb9ecc7f574c5"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.471674 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" event={"ID":"de6a04c8-bd9a-4563-85db-e83ba104deb2","Type":"ContainerStarted","Data":"1b1c27ee07e140a8373d8ca44651bfa9c4c614f2f7b5a0ac6f6f15bccf761e15"} Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.508390 4970 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-c925s container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.508658 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.509153 4970 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lr6pb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.509176 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.561536 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.562770 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.062755263 +0000 UTC m=+144.350512556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.666059 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.666409 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.166392566 +0000 UTC m=+144.454149919 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.739946 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mv65q"] Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.778196 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.778498 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.278480245 +0000 UTC m=+144.566237528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.783039 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.783414 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.283403912 +0000 UTC m=+144.571161195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.884162 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.884734 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.384718785 +0000 UTC m=+144.672476078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.933388 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb"] Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.950927 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j"] Nov 24 13:48:48 crc kubenswrapper[4970]: I1124 13:48:48.990184 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:48 crc kubenswrapper[4970]: E1124 13:48:48.990448 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.49043678 +0000 UTC m=+144.778194073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: W1124 13:48:48.994021 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b2d5731_75aa_457d_8afb_75815fd5ca93.slice/crio-9e11d2957983fb7e0a8df75f981faa01a7b809e0e967f408c8bba5704c06cd7c WatchSource:0}: Error finding container 9e11d2957983fb7e0a8df75f981faa01a7b809e0e967f408c8bba5704c06cd7c: Status 404 returned error can't find the container with id 9e11d2957983fb7e0a8df75f981faa01a7b809e0e967f408c8bba5704c06cd7c Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:48.996896 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.006944 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.012264 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.023919 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.095542 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.095658 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.595636753 +0000 UTC m=+144.883394046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.095788 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.096084 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.596073075 +0000 UTC m=+144.883830368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.096529 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" podStartSLOduration=120.096511696 podStartE2EDuration="2m0.096511696s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.095269794 +0000 UTC m=+144.383027077" watchObservedRunningTime="2025-11-24 13:48:49.096511696 +0000 UTC m=+144.384268989" Nov 24 13:48:49 crc kubenswrapper[4970]: W1124 13:48:49.124873 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d68406f_842a_4068_97e1_8efa425c71be.slice/crio-1c2e57d718da1038ef7059305d942d10107fbf52003b51e88ae21181da59ef7a WatchSource:0}: Error finding container 1c2e57d718da1038ef7059305d942d10107fbf52003b51e88ae21181da59ef7a: Status 404 returned error can't find the container with id 1c2e57d718da1038ef7059305d942d10107fbf52003b51e88ae21181da59ef7a Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.135032 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.137186 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-nhtn4"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.146997 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" podStartSLOduration=120.146971232 podStartE2EDuration="2m0.146971232s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.141408519 +0000 UTC m=+144.429165812" watchObservedRunningTime="2025-11-24 13:48:49.146971232 +0000 UTC m=+144.434728535" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.199601 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.199955 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.699941333 +0000 UTC m=+144.987698616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.219130 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" podStartSLOduration=120.219110145 podStartE2EDuration="2m0.219110145s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.216437346 +0000 UTC m=+144.504194639" watchObservedRunningTime="2025-11-24 13:48:49.219110145 +0000 UTC m=+144.506867438" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.225879 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-s97wf"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.240616 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-fsgxr" podStartSLOduration=120.240597988 podStartE2EDuration="2m0.240597988s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.238975106 +0000 UTC m=+144.526732409" watchObservedRunningTime="2025-11-24 13:48:49.240597988 +0000 UTC m=+144.528355281" Nov 24 13:48:49 crc kubenswrapper[4970]: W1124 13:48:49.247036 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0819c483_76c3_48b7_9504_b8bdaeced5a5.slice/crio-04d2a53414b471ca4c9010c0c20ff916228b81f1752d509150bde67c15a111b9 WatchSource:0}: Error finding container 04d2a53414b471ca4c9010c0c20ff916228b81f1752d509150bde67c15a111b9: Status 404 returned error can't find the container with id 04d2a53414b471ca4c9010c0c20ff916228b81f1752d509150bde67c15a111b9 Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.263145 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.283053 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tfg84" podStartSLOduration=120.283023497 podStartE2EDuration="2m0.283023497s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.268224807 +0000 UTC m=+144.555982100" watchObservedRunningTime="2025-11-24 13:48:49.283023497 +0000 UTC m=+144.570780790" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.305815 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.306161 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.806147102 +0000 UTC m=+145.093904395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.308046 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p579k" podStartSLOduration=120.30802494 podStartE2EDuration="2m0.30802494s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.293910597 +0000 UTC m=+144.581667890" watchObservedRunningTime="2025-11-24 13:48:49.30802494 +0000 UTC m=+144.595782233" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.358995 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-kxz9f"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.359229 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.367889 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.381664 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.408245 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.408801 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:49.908781378 +0000 UTC m=+145.196538671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.417343 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rsrgq"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.423901 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.443281 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vf658"] Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.498042 4970 generic.go:334] "Generic (PLEG): container finished" podID="289231a8-1bd6-4f94-b88f-6f4273e242bc" containerID="98b73bbb2844d665a654b65c4e5e9433dfaf2b37c7b117a1e78e52c897eeda7d" exitCode=0 Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.498100 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" event={"ID":"289231a8-1bd6-4f94-b88f-6f4273e242bc","Type":"ContainerDied","Data":"98b73bbb2844d665a654b65c4e5e9433dfaf2b37c7b117a1e78e52c897eeda7d"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.500170 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" event={"ID":"8d68406f-842a-4068-97e1-8efa425c71be","Type":"ContainerStarted","Data":"1c2e57d718da1038ef7059305d942d10107fbf52003b51e88ae21181da59ef7a"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.502426 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" event={"ID":"5fda55b3-c70d-40cd-8c4d-69d838eb733f","Type":"ContainerStarted","Data":"494bb858516f46777ac01aae2155b123cad3cf2b3bbe96741979d2fa44140882"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.502453 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" event={"ID":"5fda55b3-c70d-40cd-8c4d-69d838eb733f","Type":"ContainerStarted","Data":"850efc99e1e984488441a1b73fb25bd81797ab3cb3fa07bab2c6efd775bc3f84"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.510033 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.510303 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.010291866 +0000 UTC m=+145.298049159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.516743 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" event={"ID":"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f","Type":"ContainerStarted","Data":"c4082b3b22ec487d1d7892abda9aa6f9e2047009659758c7b2651c9d3f4d7e94"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.516789 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" event={"ID":"cfce80e3-ac91-47e9-b1f4-ba57d3f5b14f","Type":"ContainerStarted","Data":"6122599f6dd6375bd6a7656f6263339103f3fa282ec3cd4dc50f3f2814c55705"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.524649 4970 generic.go:334] "Generic (PLEG): container finished" podID="835318f4-7393-4594-baab-55d2ab371714" containerID="8fad0592b26d058f52f46b03ff6a23544232018e2318ecab06b9f79cbd7ba5b1" exitCode=0 Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.525765 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" event={"ID":"835318f4-7393-4594-baab-55d2ab371714","Type":"ContainerDied","Data":"8fad0592b26d058f52f46b03ff6a23544232018e2318ecab06b9f79cbd7ba5b1"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.533299 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" event={"ID":"851664d5-b007-44c4-b4f4-99bcc0dd6054","Type":"ContainerStarted","Data":"f92c4bfaa48a8476910b15ac950bd13e543b32534aa0339166605facd8f5eb85"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.537353 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tpqrz" event={"ID":"110f91cd-900b-4698-8b85-0dfa8d00cf57","Type":"ContainerStarted","Data":"ea43f4bcf4bc172d21a5fc53162800069cf91a159655ec182f6a7399f20a43dd"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.540602 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" event={"ID":"4b2d5731-75aa-457d-8afb-75815fd5ca93","Type":"ContainerStarted","Data":"9e11d2957983fb7e0a8df75f981faa01a7b809e0e967f408c8bba5704c06cd7c"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.546657 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" event={"ID":"97b2415c-4b75-495f-bf2d-8ab1a5a23cec","Type":"ContainerStarted","Data":"cdb1a9f2c373568a7db9781a91b20ef1470e26d2f756d234da6e6d17e3f9b2ec"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.546693 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" event={"ID":"97b2415c-4b75-495f-bf2d-8ab1a5a23cec","Type":"ContainerStarted","Data":"8a2d1ea9a0c0b20e90b4ece67126c2c23510bbb3b91e50ed1bf5d2664888a438"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.556688 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" event={"ID":"c0a47ee9-0bab-4235-9ca4-4088ffc0241a","Type":"ContainerStarted","Data":"e02e09981d14bb4859275849706c38f178b3c70d6a687412b7187796461d2d10"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.556754 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" event={"ID":"c0a47ee9-0bab-4235-9ca4-4088ffc0241a","Type":"ContainerStarted","Data":"10dc92b135af5cf85d0caa823115f7875d903a9aa7afeb75d991686753cc77d6"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.569403 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" event={"ID":"8f376ff7-d945-45e1-b6ff-df9769b1ecb4","Type":"ContainerStarted","Data":"fcf2c0471cc7a575cb10a4131549ad8a819ea289bf54cb0613908290aa443a9e"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.585170 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lql7t" event={"ID":"a9761a4c-d28b-475d-a6d8-8f78c29c42f6","Type":"ContainerStarted","Data":"fb5221d0dbfc0180add40e829b47b70ba7ed927e5f6005dbf6275e871055f301"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.596014 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-sxctt" podStartSLOduration=120.595997928 podStartE2EDuration="2m0.595997928s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.572426463 +0000 UTC m=+144.860183756" watchObservedRunningTime="2025-11-24 13:48:49.595997928 +0000 UTC m=+144.883755211" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.596193 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mv65q" podStartSLOduration=120.596187943 podStartE2EDuration="2m0.596187943s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.594002266 +0000 UTC m=+144.881759559" watchObservedRunningTime="2025-11-24 13:48:49.596187943 +0000 UTC m=+144.883945236" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.597515 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" event={"ID":"9ba805cc-d03e-4b6a-9ae7-1392d4253730","Type":"ContainerStarted","Data":"1fabd7c368e494a85fa7b0daff02b696b97a10111e9102727241173b3e6056a1"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.611360 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.613228 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" event={"ID":"0335a407-86a1-4ff0-9e46-df9f7ce8f3cc","Type":"ContainerStarted","Data":"0c8b6a8a855605d51cf75697b4645d9d93238595a9df1a7589fc00348a4b6518"} Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.615846 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.115823978 +0000 UTC m=+145.403581271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.621184 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-v8mhc" podStartSLOduration=120.621166435 podStartE2EDuration="2m0.621166435s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.619948994 +0000 UTC m=+144.907706287" watchObservedRunningTime="2025-11-24 13:48:49.621166435 +0000 UTC m=+144.908923728" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.624543 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" event={"ID":"6de7023f-9f6a-471b-9f6f-650e9052b9e0","Type":"ContainerStarted","Data":"407ae05faacc602031c1ab6756dae88362c432263fe7fa3e843b93db38d99b91"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.648842 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j7dr6" podStartSLOduration=120.648824995 podStartE2EDuration="2m0.648824995s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.64824376 +0000 UTC m=+144.936001053" watchObservedRunningTime="2025-11-24 13:48:49.648824995 +0000 UTC m=+144.936582288" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.669062 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-tpqrz" podStartSLOduration=5.669029214 podStartE2EDuration="5.669029214s" podCreationTimestamp="2025-11-24 13:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.661742527 +0000 UTC m=+144.949499820" watchObservedRunningTime="2025-11-24 13:48:49.669029214 +0000 UTC m=+144.956786507" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.673920 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" event={"ID":"3999597f-f349-4448-b4f4-45fa925f3342","Type":"ContainerStarted","Data":"31ad24cc5daf352c5e69497f15300cadf6a11424ffb3c61f8bc399e9c46bf38a"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.674099 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" event={"ID":"3999597f-f349-4448-b4f4-45fa925f3342","Type":"ContainerStarted","Data":"9d96c3b310133298369ec3b15ac3757193bfcb2a64d96a615803f22843e66e6f"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.676479 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nhtn4" event={"ID":"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0","Type":"ContainerStarted","Data":"1debd52d14edef10f0bb628da6ea8e264c7d4bf91fc1e6733a8bb0660c50ec7f"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.685901 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qz272" podStartSLOduration=120.68588419700001 podStartE2EDuration="2m0.685884197s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.685331433 +0000 UTC m=+144.973088726" watchObservedRunningTime="2025-11-24 13:48:49.685884197 +0000 UTC m=+144.973641490" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.691042 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" event={"ID":"8b04d05d-1031-4197-8fb3-59c52252bad7","Type":"ContainerStarted","Data":"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.691965 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.700495 4970 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m968g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.700685 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.702602 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" event={"ID":"f360574b-9321-4007-a976-36fb4ac6d812","Type":"ContainerStarted","Data":"fabfa72d92826feaf4b0c979917548c1d07ca6133b4c62014d74a171a650a5ac"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.709327 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lql7t" podStartSLOduration=120.709303109 podStartE2EDuration="2m0.709303109s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.708735044 +0000 UTC m=+144.996492337" watchObservedRunningTime="2025-11-24 13:48:49.709303109 +0000 UTC m=+144.997060402" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.710385 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" event={"ID":"d38364f3-e04f-4789-81c8-23ea7e98d39e","Type":"ContainerStarted","Data":"717062ff63b1fac08845b6742661c10b1d7939a957d83c67bc2de0e010df9aac"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.713453 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.714974 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" event={"ID":"95d62494-ded0-4f72-b037-d8f47de79dd2","Type":"ContainerStarted","Data":"7d4968a00113b87ac016fd9b01ad9d9669bc11d26c96136bd4eb0e43b0c38fbc"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.715007 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" event={"ID":"95d62494-ded0-4f72-b037-d8f47de79dd2","Type":"ContainerStarted","Data":"669b15997d71909819bd12ce65f25a1ba7bf61ddf554087a2feae7c9a778c3c3"} Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.715053 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.215040877 +0000 UTC m=+145.502798170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.729724 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8pltx" event={"ID":"550abb24-6e1a-4f62-a506-ebb047eb6b92","Type":"ContainerStarted","Data":"1692c06d692b42a355b764928d7005521e7786783675ebadde1d7bdc94c69091"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.731665 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.745740 4970 patch_prober.go:28] interesting pod/console-operator-58897d9998-8pltx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.745802 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8pltx" podUID="550abb24-6e1a-4f62-a506-ebb047eb6b92" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.746231 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" event={"ID":"a71e8ef4-11e9-4d33-9a39-256022c7c008","Type":"ContainerStarted","Data":"8ee28bb2fd0a62ab03b150b87497543c9482d42bad0c63bba6a2148a25e681cf"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.746444 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5jd8l" podStartSLOduration=120.746425063 podStartE2EDuration="2m0.746425063s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.73076145 +0000 UTC m=+145.018518743" watchObservedRunningTime="2025-11-24 13:48:49.746425063 +0000 UTC m=+145.034182366" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.748431 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" podStartSLOduration=120.748424293 podStartE2EDuration="2m0.748424293s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.746904775 +0000 UTC m=+145.034662068" watchObservedRunningTime="2025-11-24 13:48:49.748424293 +0000 UTC m=+145.036181586" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.754827 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" event={"ID":"0819c483-76c3-48b7-9504-b8bdaeced5a5","Type":"ContainerStarted","Data":"04d2a53414b471ca4c9010c0c20ff916228b81f1752d509150bde67c15a111b9"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.768253 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8pltx" podStartSLOduration=120.768232863 podStartE2EDuration="2m0.768232863s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.768087139 +0000 UTC m=+145.055844432" watchObservedRunningTime="2025-11-24 13:48:49.768232863 +0000 UTC m=+145.055990156" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.786681 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" event={"ID":"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5","Type":"ContainerStarted","Data":"b02703517fb79fd56e899ddfe9fa07cff05c008ce6d8bbeb2e6adbec9c5f8ed3"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.789672 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" event={"ID":"53b7c8f1-519f-4063-84e6-4e0b85fe57f9","Type":"ContainerStarted","Data":"598e080acff7bb25a00964a0c69d1598ad7a915e626258cca52b9c36e42752ac"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.800641 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-5s9hm" event={"ID":"f583f773-9909-4b44-8c08-f4fbd14830be","Type":"ContainerStarted","Data":"afb4dbedb364d2dd07e82008d9fed90ac3eafd9aa281aaf39f72f88ff2232af2"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.801370 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.815079 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.816357 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.316337579 +0000 UTC m=+145.604094862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.825957 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" event={"ID":"5a222693-bf87-4da6-aff7-5b13c6234a73","Type":"ContainerStarted","Data":"31c029efd89bceb0f994cacc0bedfd0c01f8eb19fc63f0a8b138468b6d70e209"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.826086 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" event={"ID":"5a222693-bf87-4da6-aff7-5b13c6234a73","Type":"ContainerStarted","Data":"5376ba172a1bf1f29d703479bc9898870bd9cb05398e522e9ea168df6b6a4882"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.826045 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.826222 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.826668 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.853371 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-t9qzb" podStartSLOduration=120.853353009 podStartE2EDuration="2m0.853353009s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.82922601 +0000 UTC m=+145.116983323" watchObservedRunningTime="2025-11-24 13:48:49.853353009 +0000 UTC m=+145.141110302" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.856409 4970 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vsj9t container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.856501 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" podUID="5a222693-bf87-4da6-aff7-5b13c6234a73" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.864252 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.889102 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:49 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:49 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:49 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.889154 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.896488 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rsrgq" event={"ID":"665b8213-0520-4614-90fa-4fc8378ce70c","Type":"ContainerStarted","Data":"e055dff4d1adb17f3908c73abefef886d949609e7de7f6367d0193a0b6b80786"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.912305 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" podStartSLOduration=120.912289073 podStartE2EDuration="2m0.912289073s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.877804028 +0000 UTC m=+145.165561321" watchObservedRunningTime="2025-11-24 13:48:49.912289073 +0000 UTC m=+145.200046366" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.913380 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-5s9hm" podStartSLOduration=120.913375091 podStartE2EDuration="2m0.913375091s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.91096903 +0000 UTC m=+145.198726323" watchObservedRunningTime="2025-11-24 13:48:49.913375091 +0000 UTC m=+145.201132374" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.917591 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:49 crc kubenswrapper[4970]: E1124 13:48:49.918052 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.418038512 +0000 UTC m=+145.705795805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.944336 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" podStartSLOduration=120.944311706 podStartE2EDuration="2m0.944311706s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.938256381 +0000 UTC m=+145.226013674" watchObservedRunningTime="2025-11-24 13:48:49.944311706 +0000 UTC m=+145.232068999" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.957918 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" event={"ID":"decc8d43-ce9c-4d67-a2a5-c4203bbc12bb","Type":"ContainerStarted","Data":"b194d9f37008854f53adce4b43007ce290a4dd89a4e478e07b50377343d51133"} Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.964834 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.967435 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:48:49 crc kubenswrapper[4970]: I1124 13:48:49.972993 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pw2s5" podStartSLOduration=120.972982383 podStartE2EDuration="2m0.972982383s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:49.971057193 +0000 UTC m=+145.258814476" watchObservedRunningTime="2025-11-24 13:48:49.972982383 +0000 UTC m=+145.260739676" Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.018463 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.019808 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.519785745 +0000 UTC m=+145.807543038 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.120144 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.120931 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.620920453 +0000 UTC m=+145.908677746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.220943 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.221291 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.721276512 +0000 UTC m=+146.009033805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.322232 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.322554 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.822543243 +0000 UTC m=+146.110300536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.423455 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.423840 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:50.923820356 +0000 UTC m=+146.211577649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.525231 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.525626 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.025611631 +0000 UTC m=+146.313368924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.626058 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.626237 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.126212385 +0000 UTC m=+146.413969678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.626606 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.626946 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.126934614 +0000 UTC m=+146.414691907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.727189 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.727610 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.22758928 +0000 UTC m=+146.515346573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.835605 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.836044 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.336024805 +0000 UTC m=+146.623782098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.876300 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:50 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:50 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:50 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.876386 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.937147 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.937280 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.437263007 +0000 UTC m=+146.725020300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:50 crc kubenswrapper[4970]: I1124 13:48:50.937419 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:50 crc kubenswrapper[4970]: E1124 13:48:50.937813 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.43780313 +0000 UTC m=+146.725560423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:50.999927 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" event={"ID":"53b7c8f1-519f-4063-84e6-4e0b85fe57f9","Type":"ContainerStarted","Data":"6fce3202ec2e0caf65c347c2f8362a4b6ef43ddb43dc94cbefa1117c443b4b12"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.036795 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" event={"ID":"a71e8ef4-11e9-4d33-9a39-256022c7c008","Type":"ContainerStarted","Data":"7b14a749c516111341a1088a0e7f6419ee3db9cd2f0c6f521cd32ff40665e690"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.037946 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.038317 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.538300392 +0000 UTC m=+146.826057685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.040101 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rsrgq" event={"ID":"665b8213-0520-4614-90fa-4fc8378ce70c","Type":"ContainerStarted","Data":"c0b60972ed67aeb0815cbfc0a8ccb241fae167cadfb04d7107e283c55e14003f"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.041596 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-nhtn4" event={"ID":"d8ff3bbd-b13b-4e33-9a26-0c48a1427fc0","Type":"ContainerStarted","Data":"0fa328f78d928d0fcaabff55848e70c48da56cb38128e66c5aeff0787abfe314"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.064857 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" event={"ID":"f360574b-9321-4007-a976-36fb4ac6d812","Type":"ContainerStarted","Data":"a1c0caa2b793df74a6f9c10b3ec6b6b905c715d277ae75e1ad3011d4dad90ca3"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.068608 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" podStartSLOduration=122.06858771 podStartE2EDuration="2m2.06858771s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.026752665 +0000 UTC m=+146.314509958" watchObservedRunningTime="2025-11-24 13:48:51.06858771 +0000 UTC m=+146.356345003" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.079757 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cnrs" event={"ID":"de4a90c4-c8e5-45f2-b8a9-419acf2d26a5","Type":"ContainerStarted","Data":"5be9c790777ad543d6171b0abeebcf8e94fed2f54ecb5ada9cec7ee180cd9da3"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.081913 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" event={"ID":"851664d5-b007-44c4-b4f4-99bcc0dd6054","Type":"ContainerStarted","Data":"c9e2ec2f23cb4f1aac493860f120892569b771f722b5fc6fa68b3dd55080a3b0"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.083538 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" event={"ID":"0819c483-76c3-48b7-9504-b8bdaeced5a5","Type":"ContainerStarted","Data":"f73140ccca9e11b31abff11688f58b7063d430b039bbce0d96ee67231e8b1ca3"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.084202 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.090220 4970 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-7x2m4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.090286 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" podUID="0819c483-76c3-48b7-9504-b8bdaeced5a5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.100208 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jhzg2" podStartSLOduration=122.100182072 podStartE2EDuration="2m2.100182072s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.070277824 +0000 UTC m=+146.358035117" watchObservedRunningTime="2025-11-24 13:48:51.100182072 +0000 UTC m=+146.387939365" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.115060 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" event={"ID":"835318f4-7393-4594-baab-55d2ab371714","Type":"ContainerStarted","Data":"67f28300535dde93baa0f1062415cd4d2e6c4620f188fc49009473819e7ea8db"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.120799 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-cxqx2" podStartSLOduration=122.120779121 podStartE2EDuration="2m2.120779121s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.101505806 +0000 UTC m=+146.389263099" watchObservedRunningTime="2025-11-24 13:48:51.120779121 +0000 UTC m=+146.408536414" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.135233 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" event={"ID":"289231a8-1bd6-4f94-b88f-6f4273e242bc","Type":"ContainerStarted","Data":"78edd4a3a6f847c28c179ed1c0a8829be02c19a38065118159ae2cf5b64bd596"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.139328 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.141327 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.641313348 +0000 UTC m=+146.929070641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.155275 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-nhtn4" podStartSLOduration=7.155254886 podStartE2EDuration="7.155254886s" podCreationTimestamp="2025-11-24 13:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.125834061 +0000 UTC m=+146.413591354" watchObservedRunningTime="2025-11-24 13:48:51.155254886 +0000 UTC m=+146.443012179" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.166626 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" event={"ID":"8d68406f-842a-4068-97e1-8efa425c71be","Type":"ContainerStarted","Data":"b75bc420ac5a77fb621ffa92d010c861a325ee5766d72b26ec99db1d0e5dcb82"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.166683 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" event={"ID":"8d68406f-842a-4068-97e1-8efa425c71be","Type":"ContainerStarted","Data":"4292675a9c2e301996dd24a88128222bcfeef636d2c07e10a39fd112f2c291a6"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.175127 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" podStartSLOduration=122.175104076 podStartE2EDuration="2m2.175104076s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.155951615 +0000 UTC m=+146.443708908" watchObservedRunningTime="2025-11-24 13:48:51.175104076 +0000 UTC m=+146.462861369" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.176362 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" podStartSLOduration=122.176352098 podStartE2EDuration="2m2.176352098s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.175196088 +0000 UTC m=+146.462953381" watchObservedRunningTime="2025-11-24 13:48:51.176352098 +0000 UTC m=+146.464109391" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.182404 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" event={"ID":"9ba805cc-d03e-4b6a-9ae7-1392d4253730","Type":"ContainerStarted","Data":"a1d397b40340d7b7df0b1a42998715a59768314dcf7ff81489fe983be7a78c50"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.196503 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-kxz9f" podStartSLOduration=122.196481796 podStartE2EDuration="2m2.196481796s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.194219758 +0000 UTC m=+146.481977051" watchObservedRunningTime="2025-11-24 13:48:51.196481796 +0000 UTC m=+146.484239099" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.197090 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" event={"ID":"5fda55b3-c70d-40cd-8c4d-69d838eb733f","Type":"ContainerStarted","Data":"45a741bb15cdd523703eafdaa646687301eb629cd9c4d4e6bb845e9007add9ba"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.208614 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.215670 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" event={"ID":"8816f5d7-5926-4600-8b05-0a0a6782ce85","Type":"ContainerStarted","Data":"5a4cf0f0ce6f633dca441679b555e1be07f70f09fa9fd4196eaf85c7bd781f36"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.215751 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" event={"ID":"8816f5d7-5926-4600-8b05-0a0a6782ce85","Type":"ContainerStarted","Data":"8a4558d88e593e160d9401210e0f30693cc12ca9292a1131cbd4532582debeca"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.217418 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.222321 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" event={"ID":"d38364f3-e04f-4789-81c8-23ea7e98d39e","Type":"ContainerStarted","Data":"424cdd9492bba6e080a9c0da0ce4a91d92072646f22e94c2c59f95bc817c4778"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.223183 4970 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-7xpcq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.223212 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" podUID="8816f5d7-5926-4600-8b05-0a0a6782ce85" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.228720 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rd2x8" podStartSLOduration=122.228698724 podStartE2EDuration="2m2.228698724s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.224032643 +0000 UTC m=+146.511789936" watchObservedRunningTime="2025-11-24 13:48:51.228698724 +0000 UTC m=+146.516456017" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.241363 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.242061 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" event={"ID":"4b2d5731-75aa-457d-8afb-75815fd5ca93","Type":"ContainerStarted","Data":"217de42cb537c2d5cf8d3bfe83441def4479e915f4a33121a34583dc6ac3f306"} Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.243041 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.743025532 +0000 UTC m=+147.030782825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.247977 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" event={"ID":"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95","Type":"ContainerStarted","Data":"503957935dbe2f225a73f273572bd43384e43fed5cba3567f6eded535d980beb"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.248184 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" event={"ID":"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95","Type":"ContainerStarted","Data":"ad4e282c82b2d7d6599117a69ebab4a2869162cd15db54369e95cf6a384d58a9"} Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249495 4970 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-m968g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249550 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249643 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249661 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249675 4970 patch_prober.go:28] interesting pod/console-operator-58897d9998-8pltx container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.249713 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8pltx" podUID="550abb24-6e1a-4f62-a506-ebb047eb6b92" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.250075 4970 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-vsj9t container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.250100 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" podUID="5a222693-bf87-4da6-aff7-5b13c6234a73" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.271898 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s259n" podStartSLOduration=122.271877563 podStartE2EDuration="2m2.271877563s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.246459049 +0000 UTC m=+146.534216342" watchObservedRunningTime="2025-11-24 13:48:51.271877563 +0000 UTC m=+146.559634856" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.305743 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" podStartSLOduration=122.305728213 podStartE2EDuration="2m2.305728213s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.272425347 +0000 UTC m=+146.560182640" watchObservedRunningTime="2025-11-24 13:48:51.305728213 +0000 UTC m=+146.593485506" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.326971 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xl78j" podStartSLOduration=122.326956308 podStartE2EDuration="2m2.326956308s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.325445609 +0000 UTC m=+146.613202902" watchObservedRunningTime="2025-11-24 13:48:51.326956308 +0000 UTC m=+146.614713601" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.327201 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" podStartSLOduration=122.327197274 podStartE2EDuration="2m2.327197274s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.306934104 +0000 UTC m=+146.594691397" watchObservedRunningTime="2025-11-24 13:48:51.327197274 +0000 UTC m=+146.614954567" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.344113 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.345557 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.845542645 +0000 UTC m=+147.133299938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.352792 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" podStartSLOduration=122.352772661 podStartE2EDuration="2m2.352772661s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.351264502 +0000 UTC m=+146.639021795" watchObservedRunningTime="2025-11-24 13:48:51.352772661 +0000 UTC m=+146.640529954" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.372377 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" podStartSLOduration=122.372362464 podStartE2EDuration="2m2.372362464s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:51.370097706 +0000 UTC m=+146.657854999" watchObservedRunningTime="2025-11-24 13:48:51.372362464 +0000 UTC m=+146.660119757" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.446073 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.446729 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:51.946714964 +0000 UTC m=+147.234472247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.550208 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.550696 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.050667295 +0000 UTC m=+147.338424588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.593637 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.593832 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.595533 4970 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hz58r container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.595601 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" podUID="835318f4-7393-4594-baab-55d2ab371714" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.651493 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.651694 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.15166775 +0000 UTC m=+147.439425043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.651816 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.652082 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.15207059 +0000 UTC m=+147.439827883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.752638 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.753015 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.252987823 +0000 UTC m=+147.540745116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.854215 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.854608 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.354590773 +0000 UTC m=+147.642348066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.865925 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:51 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:51 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:51 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.865998 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:51 crc kubenswrapper[4970]: I1124 13:48:51.955698 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:51 crc kubenswrapper[4970]: E1124 13:48:51.956136 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.455932897 +0000 UTC m=+147.743690200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.057507 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.057808 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.557797734 +0000 UTC m=+147.845555017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.158244 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.158409 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.658388468 +0000 UTC m=+147.946145761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.158615 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.158896 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.658885211 +0000 UTC m=+147.946642494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.163300 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-72mjc" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.253161 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rsrgq" event={"ID":"665b8213-0520-4614-90fa-4fc8378ce70c","Type":"ContainerStarted","Data":"21251951bac1019026697eb5071be05a7a3f4149a74ae1567a8d2c68657a9cc7"} Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.253243 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rsrgq" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.254675 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ps7l4" event={"ID":"d38364f3-e04f-4789-81c8-23ea7e98d39e","Type":"ContainerStarted","Data":"5ae4dc41aada7e39247eff34cda4a91357dfe0c8943edeb2eea6ef81dafec4a4"} Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.256245 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" event={"ID":"6de7023f-9f6a-471b-9f6f-650e9052b9e0","Type":"ContainerStarted","Data":"662a1fb1e7338689a17b0043ee6d6a25e119e73746ff5be91f2b16935cc01f36"} Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.257876 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vf658" event={"ID":"611aaa5a-fc7a-43d3-a4b4-5ac2928cde95","Type":"ContainerStarted","Data":"74076fa22639e257db883f19a82abbc9a4eaef22f68ad32510e24fada404793b"} Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.259057 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.259236 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.759216739 +0000 UTC m=+148.046974032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.259293 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.259613 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.759602818 +0000 UTC m=+148.047360171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.261709 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" event={"ID":"289231a8-1bd6-4f94-b88f-6f4273e242bc","Type":"ContainerStarted","Data":"ba667858e1d50772b998af6e98a52a9f72832bf974db4fbcf42aebba844cf718"} Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.262554 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.262596 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.266838 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.272263 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-vsj9t" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.275528 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7xpcq" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.278561 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rsrgq" podStartSLOduration=8.278548104 podStartE2EDuration="8.278548104s" podCreationTimestamp="2025-11-24 13:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:52.278353269 +0000 UTC m=+147.566110562" watchObservedRunningTime="2025-11-24 13:48:52.278548104 +0000 UTC m=+147.566305397" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.360005 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.360149 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.860123801 +0000 UTC m=+148.147881094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.360258 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.365964 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.366437 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.366854 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.367072 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.370287 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.370735 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.372742 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.872726175 +0000 UTC m=+148.160483468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.375215 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.375731 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.393860 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.397631 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" podStartSLOduration=123.397616094 podStartE2EDuration="2m3.397616094s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:52.395766816 +0000 UTC m=+147.683524109" watchObservedRunningTime="2025-11-24 13:48:52.397616094 +0000 UTC m=+147.685373387" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.406602 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.468514 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.468792 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:52.968777772 +0000 UTC m=+148.256535065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.569638 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.570056 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.070040234 +0000 UTC m=+148.357797527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.608150 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.646491 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8pltx" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.651760 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-7x2m4" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.683082 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.683402 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.183388446 +0000 UTC m=+148.471145739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.790119 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.790404 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.290393885 +0000 UTC m=+148.578151178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.883187 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:52 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:52 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:52 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.891592 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.892014 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.892329 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.392312733 +0000 UTC m=+148.680070026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:52 crc kubenswrapper[4970]: I1124 13:48:52.993347 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:52 crc kubenswrapper[4970]: E1124 13:48:52.993664 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.493652466 +0000 UTC m=+148.781409759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.095014 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.095400 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.5953831 +0000 UTC m=+148.883140383 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.170337 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.171165 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.173910 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.196028 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.196394 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.696381325 +0000 UTC m=+148.984138618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.271604 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.297754 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"bef86be9d8c9f0d3d70c1def364a7dc6d187c33ff2c5a1ae427abcb1d87fff2e"} Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.298506 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.298676 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.298695 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.298740 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktl45\" (UniqueName: \"kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.298843 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.798830047 +0000 UTC m=+149.086587340 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.310257 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"590b634b702c977c3af8db5cafa5deaa3e011535966b5f1010fba4679c10ef51"} Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.326839 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" event={"ID":"6de7023f-9f6a-471b-9f6f-650e9052b9e0","Type":"ContainerStarted","Data":"af1691554e87c27af3bea19e2ef05faa6ab020de1638f22032cdb5c9df3a26d1"} Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.362307 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.363164 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.366030 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.377012 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.400117 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.400178 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.400202 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.400249 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktl45\" (UniqueName: \"kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.400795 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:53.900779516 +0000 UTC m=+149.188536799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.401220 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.401426 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.428619 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktl45\" (UniqueName: \"kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45\") pod \"community-operators-9b6jr\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.501046 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.501618 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpxhl\" (UniqueName: \"kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.501686 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.501735 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.502432 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.002415717 +0000 UTC m=+149.290173010 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.511673 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.560959 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.562082 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.567250 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.603107 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.603182 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpxhl\" (UniqueName: \"kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.603209 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.603232 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.603631 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.604066 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.104054838 +0000 UTC m=+149.391812121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.604562 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.621265 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpxhl\" (UniqueName: \"kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl\") pod \"certified-operators-h4vlc\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.648957 4970 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.687161 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.704707 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.704885 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.204859768 +0000 UTC m=+149.492617061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.704985 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.705018 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5tsl\" (UniqueName: \"kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.705045 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.705092 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.705403 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.205391692 +0000 UTC m=+149.493148985 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.760819 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.762424 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.806787 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.807231 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.807259 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5tsl\" (UniqueName: \"kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.807310 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.807670 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.807728 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.307715131 +0000 UTC m=+149.595472424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.807918 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.808124 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.809087 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:48:53 crc kubenswrapper[4970]: W1124 13:48:53.817145 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf026ffe7_2203_429b_9144_e96b21b38a7f.slice/crio-f6cd24614edc0e68ce89e52984475d91b89e7dfc81dbd90ad33ba922a25fc7cb WatchSource:0}: Error finding container f6cd24614edc0e68ce89e52984475d91b89e7dfc81dbd90ad33ba922a25fc7cb: Status 404 returned error can't find the container with id f6cd24614edc0e68ce89e52984475d91b89e7dfc81dbd90ad33ba922a25fc7cb Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.838202 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5tsl\" (UniqueName: \"kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl\") pod \"community-operators-49x5r\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.868776 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:53 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:53 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:53 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.868834 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.908727 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.908764 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.908793 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.908816 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwkq\" (UniqueName: \"kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:53 crc kubenswrapper[4970]: E1124 13:48:53.909071 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.409055904 +0000 UTC m=+149.696813197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:53 crc kubenswrapper[4970]: I1124 13:48:53.923376 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.004625 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.010298 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.010480 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.510454259 +0000 UTC m=+149.798211542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.010637 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.010681 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.010750 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.011059 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.511044504 +0000 UTC m=+149.798801797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.011093 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.011163 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwkq\" (UniqueName: \"kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.011196 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.028766 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwkq\" (UniqueName: \"kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq\") pod \"certified-operators-c9jhs\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.113139 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.113504 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.613476046 +0000 UTC m=+149.901233339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.113809 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.114137 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.614122112 +0000 UTC m=+149.901879415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.121308 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.215300 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.215604 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.715590359 +0000 UTC m=+150.003347652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.261910 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:48:54 crc kubenswrapper[4970]: W1124 13:48:54.270138 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod238eae5b_0891_4035_bd9d_a427bc7d713d.slice/crio-7c46bf27ecfd5c669757457452fd9097d4cdf62511c27843ac3e84589ac0dabf WatchSource:0}: Error finding container 7c46bf27ecfd5c669757457452fd9097d4cdf62511c27843ac3e84589ac0dabf: Status 404 returned error can't find the container with id 7c46bf27ecfd5c669757457452fd9097d4cdf62511c27843ac3e84589ac0dabf Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.318382 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.318857 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.818836832 +0000 UTC m=+150.106594135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.334785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fc1e68a5d6357c3963c0b9b70231118e6e60ed39ce37fc3a89caee77b85dc1e0"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.337180 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9b7fc609f31c831d0dbcbcbd4697224bbbef981c1f56c1da417907e6057d8140"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.337518 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.345432 4970 generic.go:334] "Generic (PLEG): container finished" podID="72e5f1c1-078d-4394-9f04-5541716da154" containerID="07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf" exitCode=0 Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.345538 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerDied","Data":"07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.345606 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerStarted","Data":"74814213c8736640ae0e82105afc001e2774ecbd07315ea17b58433c88ecf745"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.347869 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.351408 4970 generic.go:334] "Generic (PLEG): container finished" podID="53b7c8f1-519f-4063-84e6-4e0b85fe57f9" containerID="6fce3202ec2e0caf65c347c2f8362a4b6ef43ddb43dc94cbefa1117c443b4b12" exitCode=0 Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.351497 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" event={"ID":"53b7c8f1-519f-4063-84e6-4e0b85fe57f9","Type":"ContainerDied","Data":"6fce3202ec2e0caf65c347c2f8362a4b6ef43ddb43dc94cbefa1117c443b4b12"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.361735 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" event={"ID":"6de7023f-9f6a-471b-9f6f-650e9052b9e0","Type":"ContainerStarted","Data":"e2e5dc6876ce4b78dd6e4dfaa7b842411a22476833c134b686861c6692f3051e"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.361788 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" event={"ID":"6de7023f-9f6a-471b-9f6f-650e9052b9e0","Type":"ContainerStarted","Data":"c0c5658052fd90077d52187374b7c06711106c6b3b3eebe1524d1f9c13cde3a9"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.364926 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0121653a8743d83041f087dc4d65928367b9b5e4f32c90778332b89d39d3f02f"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.364991 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"824ce2c1cf0286f7735a71047852e61721a91612dca92a8f1dffefd99bbd0506"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.365107 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.367536 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerStarted","Data":"7c46bf27ecfd5c669757457452fd9097d4cdf62511c27843ac3e84589ac0dabf"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.369910 4970 generic.go:334] "Generic (PLEG): container finished" podID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerID="bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee" exitCode=0 Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.370726 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerDied","Data":"bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee"} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.370785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerStarted","Data":"f6cd24614edc0e68ce89e52984475d91b89e7dfc81dbd90ad33ba922a25fc7cb"} Nov 24 13:48:54 crc kubenswrapper[4970]: W1124 13:48:54.377707 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod973d682e_c2a5_48ed_9787_7f38271e3a6d.slice/crio-c74fe18dadb36a317788c15f2c33095402700d6ff3c246ac268747bbb9e69ade WatchSource:0}: Error finding container c74fe18dadb36a317788c15f2c33095402700d6ff3c246ac268747bbb9e69ade: Status 404 returned error can't find the container with id c74fe18dadb36a317788c15f2c33095402700d6ff3c246ac268747bbb9e69ade Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.419790 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.420480 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:54.920453442 +0000 UTC m=+150.208210735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.454143 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-s97wf" podStartSLOduration=10.454122897 podStartE2EDuration="10.454122897s" podCreationTimestamp="2025-11-24 13:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:54.453184703 +0000 UTC m=+149.740941996" watchObservedRunningTime="2025-11-24 13:48:54.454122897 +0000 UTC m=+149.741880190" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.521546 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.521845 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:55.021823177 +0000 UTC m=+150.309580470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.623362 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.623536 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 13:48:55.123512509 +0000 UTC m=+150.411269802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.623697 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: E1124 13:48:54.624041 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 13:48:55.124026792 +0000 UTC m=+150.411784095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rgct9" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.648250 4970 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T13:48:53.648978672Z","Handler":null,"Name":""} Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.652309 4970 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.652361 4970 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.724874 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.735571 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.826412 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.829249 4970 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.829285 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.859261 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rgct9\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.866965 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:54 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:54 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:54 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.867031 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:54 crc kubenswrapper[4970]: I1124 13:48:54.968096 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.155019 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.155931 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.159592 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.167671 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.334794 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.334850 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.334947 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwslw\" (UniqueName: \"kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.377029 4970 generic.go:334] "Generic (PLEG): container finished" podID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerID="8fb20d4cbdb416c26d27ce5110a9f701a5d8db0ee607ce464fbc310cf46db546" exitCode=0 Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.377279 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerDied","Data":"8fb20d4cbdb416c26d27ce5110a9f701a5d8db0ee607ce464fbc310cf46db546"} Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.377677 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerStarted","Data":"c74fe18dadb36a317788c15f2c33095402700d6ff3c246ac268747bbb9e69ade"} Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.388927 4970 generic.go:334] "Generic (PLEG): container finished" podID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerID="dd2235cb2a09b7f1055b9db2fd6319bcbbdcac29f582b6ab9c9311134ec6f096" exitCode=0 Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.389041 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerDied","Data":"dd2235cb2a09b7f1055b9db2fd6319bcbbdcac29f582b6ab9c9311134ec6f096"} Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.415729 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.435653 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.435715 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.435758 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwslw\" (UniqueName: \"kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.436874 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.437096 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: W1124 13:48:55.438137 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d5d2b6d_30d7_4504_8fb8_b11e7eb576d8.slice/crio-8287ef6534faf9eb3d568fcb4f2e9b94d37ee1cb04d3a0e048f7f46989d46889 WatchSource:0}: Error finding container 8287ef6534faf9eb3d568fcb4f2e9b94d37ee1cb04d3a0e048f7f46989d46889: Status 404 returned error can't find the container with id 8287ef6534faf9eb3d568fcb4f2e9b94d37ee1cb04d3a0e048f7f46989d46889 Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.453386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwslw\" (UniqueName: \"kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw\") pod \"redhat-marketplace-9pgnr\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.485564 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.486086 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.557926 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.559174 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.575605 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.625049 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.625877 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.629337 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.629366 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.646226 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.698834 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.742120 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.742363 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.742523 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld6jk\" (UniqueName: \"kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.742557 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.742640 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.843816 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume\") pod \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844205 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktg4h\" (UniqueName: \"kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h\") pod \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844237 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume\") pod \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\" (UID: \"53b7c8f1-519f-4063-84e6-4e0b85fe57f9\") " Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844455 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld6jk\" (UniqueName: \"kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844484 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844520 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844555 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.844641 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.845064 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.846144 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.846757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume" (OuterVolumeSpecName: "config-volume") pod "53b7c8f1-519f-4063-84e6-4e0b85fe57f9" (UID: "53b7c8f1-519f-4063-84e6-4e0b85fe57f9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.847054 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.857186 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "53b7c8f1-519f-4063-84e6-4e0b85fe57f9" (UID: "53b7c8f1-519f-4063-84e6-4e0b85fe57f9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.866468 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.868131 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:55 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:55 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:55 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.868186 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.868913 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h" (OuterVolumeSpecName: "kube-api-access-ktg4h") pod "53b7c8f1-519f-4063-84e6-4e0b85fe57f9" (UID: "53b7c8f1-519f-4063-84e6-4e0b85fe57f9"). InnerVolumeSpecName "kube-api-access-ktg4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.869814 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld6jk\" (UniqueName: \"kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk\") pod \"redhat-marketplace-szhpq\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.904767 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.945549 4970 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.945569 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktg4h\" (UniqueName: \"kubernetes.io/projected/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-kube-api-access-ktg4h\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.945597 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/53b7c8f1-519f-4063-84e6-4e0b85fe57f9-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.948917 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:48:55 crc kubenswrapper[4970]: I1124 13:48:55.959667 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:48:55 crc kubenswrapper[4970]: W1124 13:48:55.968877 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0fe399e_973f_4c4c_8e49_3619c8788ed2.slice/crio-58ecc9cc6be157f94df1e49b2735bc04e1d3aacfccd207f191a834325824ab7c WatchSource:0}: Error finding container 58ecc9cc6be157f94df1e49b2735bc04e1d3aacfccd207f191a834325824ab7c: Status 404 returned error can't find the container with id 58ecc9cc6be157f94df1e49b2735bc04e1d3aacfccd207f191a834325824ab7c Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.148079 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.148544 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.151066 4970 patch_prober.go:28] interesting pod/console-f9d7485db-fsgxr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.151120 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-fsgxr" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.183629 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.196482 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 13:48:56 crc kubenswrapper[4970]: W1124 13:48:56.204265 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd809518e_0ed6_43ee_b375_8c63acd0584e.slice/crio-65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647 WatchSource:0}: Error finding container 65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647: Status 404 returned error can't find the container with id 65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647 Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.403262 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" event={"ID":"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8","Type":"ContainerStarted","Data":"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.403635 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" event={"ID":"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8","Type":"ContainerStarted","Data":"8287ef6534faf9eb3d568fcb4f2e9b94d37ee1cb04d3a0e048f7f46989d46889"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.403931 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.407600 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" event={"ID":"53b7c8f1-519f-4063-84e6-4e0b85fe57f9","Type":"ContainerDied","Data":"598e080acff7bb25a00964a0c69d1598ad7a915e626258cca52b9c36e42752ac"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.407657 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="598e080acff7bb25a00964a0c69d1598ad7a915e626258cca52b9c36e42752ac" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.407692 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.411186 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerStarted","Data":"80daffd856772ce6a7708f4479d95062c210a0e1b452fd6d878f83a7fe7092ae"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.428028 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" podStartSLOduration=127.428010208 podStartE2EDuration="2m7.428010208s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:48:56.422729522 +0000 UTC m=+151.710486815" watchObservedRunningTime="2025-11-24 13:48:56.428010208 +0000 UTC m=+151.715767501" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.440530 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerID="f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77" exitCode=0 Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.440621 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerDied","Data":"f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.440700 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerStarted","Data":"58ecc9cc6be157f94df1e49b2735bc04e1d3aacfccd207f191a834325824ab7c"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.443924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d809518e-0ed6-43ee-b375-8c63acd0584e","Type":"ContainerStarted","Data":"65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647"} Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.469300 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.469355 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.469553 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.469658 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.554903 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:48:56 crc kubenswrapper[4970]: E1124 13:48:56.555119 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b7c8f1-519f-4063-84e6-4e0b85fe57f9" containerName="collect-profiles" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.555130 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b7c8f1-519f-4063-84e6-4e0b85fe57f9" containerName="collect-profiles" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.555242 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b7c8f1-519f-4063-84e6-4e0b85fe57f9" containerName="collect-profiles" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.559305 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.563312 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.565135 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.598889 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.604394 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hz58r" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.619295 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.619334 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.625977 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.658228 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbxs8\" (UniqueName: \"kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.658365 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.658431 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.761137 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.761239 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.761294 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbxs8\" (UniqueName: \"kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.762670 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.762712 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.801805 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbxs8\" (UniqueName: \"kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8\") pod \"redhat-operators-j24dc\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.865646 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.866734 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:56 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:56 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:56 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.866763 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.888201 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.963147 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.964812 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:56 crc kubenswrapper[4970]: I1124 13:48:56.972648 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.065705 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f89x\" (UniqueName: \"kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.065747 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.065768 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.166841 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f89x\" (UniqueName: \"kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.166892 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.166914 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.167259 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.169897 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.170021 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: W1124 13:48:57.186076 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84c15935_c1f3_4faf_b2a6_a122603f15c7.slice/crio-62ea9a79f63d6ed14c4fb8502442754547d457ecdf189b43368dc223ddfbb94e WatchSource:0}: Error finding container 62ea9a79f63d6ed14c4fb8502442754547d457ecdf189b43368dc223ddfbb94e: Status 404 returned error can't find the container with id 62ea9a79f63d6ed14c4fb8502442754547d457ecdf189b43368dc223ddfbb94e Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.194746 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f89x\" (UniqueName: \"kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x\") pod \"redhat-operators-n8rxn\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.291892 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.452933 4970 generic.go:334] "Generic (PLEG): container finished" podID="d809518e-0ed6-43ee-b375-8c63acd0584e" containerID="781c39da233c5edb3e805501083af0b960c2c21d95708a1e3521704b3574b968" exitCode=0 Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.453585 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d809518e-0ed6-43ee-b375-8c63acd0584e","Type":"ContainerDied","Data":"781c39da233c5edb3e805501083af0b960c2c21d95708a1e3521704b3574b968"} Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.483565 4970 generic.go:334] "Generic (PLEG): container finished" podID="938f9e13-38a8-45db-9e07-1258ced9d931" containerID="da7934b9d7ce4b10dec3c85b43e641a751e62c6f0f7b8829b65ae3d3fa9dbec5" exitCode=0 Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.491134 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerDied","Data":"da7934b9d7ce4b10dec3c85b43e641a751e62c6f0f7b8829b65ae3d3fa9dbec5"} Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.528421 4970 generic.go:334] "Generic (PLEG): container finished" podID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerID="3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c" exitCode=0 Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.528812 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerDied","Data":"3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c"} Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.529495 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerStarted","Data":"62ea9a79f63d6ed14c4fb8502442754547d457ecdf189b43368dc223ddfbb94e"} Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.533256 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-d49vj" Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.725997 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:48:57 crc kubenswrapper[4970]: W1124 13:48:57.749699 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3abc2b3_c6fd_41cd_904c_cdbff1044c5e.slice/crio-88a1784032d8ed89f97ade895d103e3fb5ee4665ffcc890fb74760c45acc5cb1 WatchSource:0}: Error finding container 88a1784032d8ed89f97ade895d103e3fb5ee4665ffcc890fb74760c45acc5cb1: Status 404 returned error can't find the container with id 88a1784032d8ed89f97ade895d103e3fb5ee4665ffcc890fb74760c45acc5cb1 Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.865770 4970 patch_prober.go:28] interesting pod/router-default-5444994796-lql7t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 13:48:57 crc kubenswrapper[4970]: [-]has-synced failed: reason withheld Nov 24 13:48:57 crc kubenswrapper[4970]: [+]process-running ok Nov 24 13:48:57 crc kubenswrapper[4970]: healthz check failed Nov 24 13:48:57 crc kubenswrapper[4970]: I1124 13:48:57.865853 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lql7t" podUID="a9761a4c-d28b-475d-a6d8-8f78c29c42f6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.526418 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.529091 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.540074 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.540127 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.550085 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.569312 4970 generic.go:334] "Generic (PLEG): container finished" podID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerID="6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8" exitCode=0 Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.569537 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerDied","Data":"6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8"} Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.569601 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerStarted","Data":"88a1784032d8ed89f97ade895d103e3fb5ee4665ffcc890fb74760c45acc5cb1"} Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.595784 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.595844 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.697218 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.697435 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.700048 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.724959 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.866931 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.867274 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:48:58 crc kubenswrapper[4970]: I1124 13:48:58.870859 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lql7t" Nov 24 13:49:00 crc kubenswrapper[4970]: I1124 13:49:00.801617 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:49:02 crc kubenswrapper[4970]: I1124 13:49:02.343808 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rsrgq" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.452820 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.599779 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access\") pod \"d809518e-0ed6-43ee-b375-8c63acd0584e\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.599842 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir\") pod \"d809518e-0ed6-43ee-b375-8c63acd0584e\" (UID: \"d809518e-0ed6-43ee-b375-8c63acd0584e\") " Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.600110 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d809518e-0ed6-43ee-b375-8c63acd0584e" (UID: "d809518e-0ed6-43ee-b375-8c63acd0584e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.606641 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d809518e-0ed6-43ee-b375-8c63acd0584e" (UID: "d809518e-0ed6-43ee-b375-8c63acd0584e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.621829 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d809518e-0ed6-43ee-b375-8c63acd0584e","Type":"ContainerDied","Data":"65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647"} Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.621872 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65795703fd6f8c10507d046233619f8afc104e7f4a7ae91cb58d14ee72c31647" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.622181 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.701563 4970 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d809518e-0ed6-43ee-b375-8c63acd0584e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:05 crc kubenswrapper[4970]: I1124 13:49:05.701799 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d809518e-0ed6-43ee-b375-8c63acd0584e-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.155541 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.160085 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.469495 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.469495 4970 patch_prober.go:28] interesting pod/downloads-7954f5f757-5s9hm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.470640 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:49:06 crc kubenswrapper[4970]: I1124 13:49:06.470612 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-5s9hm" podUID="f583f773-9909-4b44-8c08-f4fbd14830be" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 24 13:49:11 crc kubenswrapper[4970]: I1124 13:49:11.204961 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:49:11 crc kubenswrapper[4970]: I1124 13:49:11.205431 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:49:11 crc kubenswrapper[4970]: I1124 13:49:11.987233 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:49:11 crc kubenswrapper[4970]: I1124 13:49:11.993407 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b0fceff-420b-4465-9afa-99af03170306-metrics-certs\") pod \"network-metrics-daemon-kv7zv\" (UID: \"8b0fceff-420b-4465-9afa-99af03170306\") " pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:49:12 crc kubenswrapper[4970]: I1124 13:49:12.094534 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kv7zv" Nov 24 13:49:14 crc kubenswrapper[4970]: I1124 13:49:14.974087 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:49:16 crc kubenswrapper[4970]: I1124 13:49:16.484797 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-5s9hm" Nov 24 13:49:20 crc kubenswrapper[4970]: E1124 13:49:20.744143 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 13:49:20 crc kubenswrapper[4970]: E1124 13:49:20.744782 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zpxhl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-h4vlc_openshift-marketplace(72e5f1c1-078d-4394-9f04-5541716da154): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:49:20 crc kubenswrapper[4970]: E1124 13:49:20.745902 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-h4vlc" podUID="72e5f1c1-078d-4394-9f04-5541716da154" Nov 24 13:49:23 crc kubenswrapper[4970]: E1124 13:49:23.425766 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-h4vlc" podUID="72e5f1c1-078d-4394-9f04-5541716da154" Nov 24 13:49:23 crc kubenswrapper[4970]: E1124 13:49:23.603257 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 13:49:23 crc kubenswrapper[4970]: E1124 13:49:23.603677 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ktl45,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9b6jr_openshift-marketplace(f026ffe7-2203-429b-9144-e96b21b38a7f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 13:49:23 crc kubenswrapper[4970]: E1124 13:49:23.604884 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9b6jr" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" Nov 24 13:49:23 crc kubenswrapper[4970]: E1124 13:49:23.739561 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9b6jr" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" Nov 24 13:49:23 crc kubenswrapper[4970]: I1124 13:49:23.838253 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 13:49:23 crc kubenswrapper[4970]: I1124 13:49:23.988560 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kv7zv"] Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.747677 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerID="b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.747825 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerDied","Data":"b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.752626 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b9d9dee2-2474-41c7-bb47-dcada989f1bd","Type":"ContainerStarted","Data":"7c05ea0e863e5334072088b9f2d454bf9710aacff9f98f31c7ff09d59ce60ab2"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.752711 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b9d9dee2-2474-41c7-bb47-dcada989f1bd","Type":"ContainerStarted","Data":"34a5fc34ea87574c9a3b2491be8b0ba8a21c317c01ffc7532a113e83ed1d9145"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.761107 4970 generic.go:334] "Generic (PLEG): container finished" podID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerID="262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.761194 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerDied","Data":"262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.764168 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" event={"ID":"8b0fceff-420b-4465-9afa-99af03170306","Type":"ContainerStarted","Data":"e97548c5cc0cff89445eb79cc5af44084e26488e551e98f7dffc15b8804c6bfd"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.764214 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" event={"ID":"8b0fceff-420b-4465-9afa-99af03170306","Type":"ContainerStarted","Data":"baaf098912680878336ae3c2c4f3add9dd6c4db77c58b49f48090f4ff7bf7d3e"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.764234 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kv7zv" event={"ID":"8b0fceff-420b-4465-9afa-99af03170306","Type":"ContainerStarted","Data":"4f9f9609d30e5e9011af1924bef4a7a2f8981bb813e60a8e4079799512c50e78"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.771231 4970 generic.go:334] "Generic (PLEG): container finished" podID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerID="d8f11710cf7a94da9a68e32f05016c6cde6ba68501cb61445b5fd1c4a1e8823f" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.771316 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerDied","Data":"d8f11710cf7a94da9a68e32f05016c6cde6ba68501cb61445b5fd1c4a1e8823f"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.780495 4970 generic.go:334] "Generic (PLEG): container finished" podID="938f9e13-38a8-45db-9e07-1258ced9d931" containerID="93f84f149b321e4246f57f45504f5d37f9768788aa54506f5d8b40f560c99da7" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.780630 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerDied","Data":"93f84f149b321e4246f57f45504f5d37f9768788aa54506f5d8b40f560c99da7"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.786938 4970 generic.go:334] "Generic (PLEG): container finished" podID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerID="71a6230f920181f5efcf73dd4fe3a007c576ef61dac5ffb91d0c1f26477c17db" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.787036 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerDied","Data":"71a6230f920181f5efcf73dd4fe3a007c576ef61dac5ffb91d0c1f26477c17db"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.790300 4970 generic.go:334] "Generic (PLEG): container finished" podID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerID="c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd" exitCode=0 Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.790401 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerDied","Data":"c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd"} Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.842145 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kv7zv" podStartSLOduration=155.842127242 podStartE2EDuration="2m35.842127242s" podCreationTimestamp="2025-11-24 13:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:49:24.840728116 +0000 UTC m=+180.128485449" watchObservedRunningTime="2025-11-24 13:49:24.842127242 +0000 UTC m=+180.129884535" Nov 24 13:49:24 crc kubenswrapper[4970]: I1124 13:49:24.869188 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=26.869159736 podStartE2EDuration="26.869159736s" podCreationTimestamp="2025-11-24 13:48:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:49:24.860022052 +0000 UTC m=+180.147779335" watchObservedRunningTime="2025-11-24 13:49:24.869159736 +0000 UTC m=+180.156917069" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.797898 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerStarted","Data":"1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.801264 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerStarted","Data":"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.805898 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerStarted","Data":"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.808275 4970 generic.go:334] "Generic (PLEG): container finished" podID="b9d9dee2-2474-41c7-bb47-dcada989f1bd" containerID="7c05ea0e863e5334072088b9f2d454bf9710aacff9f98f31c7ff09d59ce60ab2" exitCode=0 Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.808343 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b9d9dee2-2474-41c7-bb47-dcada989f1bd","Type":"ContainerDied","Data":"7c05ea0e863e5334072088b9f2d454bf9710aacff9f98f31c7ff09d59ce60ab2"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.810616 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerStarted","Data":"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.815171 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerStarted","Data":"75578512a5f05ee89528934a829e0627d705ac5d5f361ebe6b1fca298a610f7b"} Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.820193 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-49x5r" podStartSLOduration=2.933814002 podStartE2EDuration="32.820172078s" podCreationTimestamp="2025-11-24 13:48:53 +0000 UTC" firstStartedPulling="2025-11-24 13:48:55.398409837 +0000 UTC m=+150.686167130" lastFinishedPulling="2025-11-24 13:49:25.284767903 +0000 UTC m=+180.572525206" observedRunningTime="2025-11-24 13:49:25.817818878 +0000 UTC m=+181.105576171" watchObservedRunningTime="2025-11-24 13:49:25.820172078 +0000 UTC m=+181.107929391" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.836430 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9pgnr" podStartSLOduration=1.928009854 podStartE2EDuration="30.836370675s" podCreationTimestamp="2025-11-24 13:48:55 +0000 UTC" firstStartedPulling="2025-11-24 13:48:56.442171792 +0000 UTC m=+151.729929105" lastFinishedPulling="2025-11-24 13:49:25.350532633 +0000 UTC m=+180.638289926" observedRunningTime="2025-11-24 13:49:25.831325634 +0000 UTC m=+181.119082927" watchObservedRunningTime="2025-11-24 13:49:25.836370675 +0000 UTC m=+181.124127968" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.877837 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n8rxn" podStartSLOduration=3.27195191 podStartE2EDuration="29.877803449s" podCreationTimestamp="2025-11-24 13:48:56 +0000 UTC" firstStartedPulling="2025-11-24 13:48:58.572194733 +0000 UTC m=+153.859952026" lastFinishedPulling="2025-11-24 13:49:25.178046262 +0000 UTC m=+180.465803565" observedRunningTime="2025-11-24 13:49:25.855143406 +0000 UTC m=+181.142900699" watchObservedRunningTime="2025-11-24 13:49:25.877803449 +0000 UTC m=+181.165560742" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.878243 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j24dc" podStartSLOduration=2.0788218880000002 podStartE2EDuration="29.87823896s" podCreationTimestamp="2025-11-24 13:48:56 +0000 UTC" firstStartedPulling="2025-11-24 13:48:57.534130605 +0000 UTC m=+152.821887898" lastFinishedPulling="2025-11-24 13:49:25.333547677 +0000 UTC m=+180.621304970" observedRunningTime="2025-11-24 13:49:25.87510453 +0000 UTC m=+181.162861823" watchObservedRunningTime="2025-11-24 13:49:25.87823896 +0000 UTC m=+181.165996253" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.894704 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-szhpq" podStartSLOduration=3.064258685 podStartE2EDuration="30.894688223s" podCreationTimestamp="2025-11-24 13:48:55 +0000 UTC" firstStartedPulling="2025-11-24 13:48:57.485399413 +0000 UTC m=+152.773156696" lastFinishedPulling="2025-11-24 13:49:25.315828931 +0000 UTC m=+180.603586234" observedRunningTime="2025-11-24 13:49:25.89223242 +0000 UTC m=+181.179989713" watchObservedRunningTime="2025-11-24 13:49:25.894688223 +0000 UTC m=+181.182445516" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.906647 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:25 crc kubenswrapper[4970]: I1124 13:49:25.906839 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:26 crc kubenswrapper[4970]: I1124 13:49:26.824312 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerStarted","Data":"b9412a5c54c6986bc5aab9b182af91c9951d52b65a5ccfa2e45826585709ee20"} Nov 24 13:49:26 crc kubenswrapper[4970]: I1124 13:49:26.841990 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c9jhs" podStartSLOduration=3.540154849 podStartE2EDuration="33.841973809s" podCreationTimestamp="2025-11-24 13:48:53 +0000 UTC" firstStartedPulling="2025-11-24 13:48:55.386939632 +0000 UTC m=+150.674696925" lastFinishedPulling="2025-11-24 13:49:25.688758592 +0000 UTC m=+180.976515885" observedRunningTime="2025-11-24 13:49:26.839080865 +0000 UTC m=+182.126838168" watchObservedRunningTime="2025-11-24 13:49:26.841973809 +0000 UTC m=+182.129731102" Nov 24 13:49:26 crc kubenswrapper[4970]: I1124 13:49:26.890861 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:49:26 crc kubenswrapper[4970]: I1124 13:49:26.891089 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:49:26 crc kubenswrapper[4970]: I1124 13:49:26.947108 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-d9kzs" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.090784 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-szhpq" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="registry-server" probeResult="failure" output=< Nov 24 13:49:27 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 13:49:27 crc kubenswrapper[4970]: > Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.225953 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.292943 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.292979 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.294690 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir\") pod \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.294780 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access\") pod \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\" (UID: \"b9d9dee2-2474-41c7-bb47-dcada989f1bd\") " Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.295065 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b9d9dee2-2474-41c7-bb47-dcada989f1bd" (UID: "b9d9dee2-2474-41c7-bb47-dcada989f1bd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.301811 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b9d9dee2-2474-41c7-bb47-dcada989f1bd" (UID: "b9d9dee2-2474-41c7-bb47-dcada989f1bd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.396186 4970 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.396230 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b9d9dee2-2474-41c7-bb47-dcada989f1bd-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.835209 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.835209 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b9d9dee2-2474-41c7-bb47-dcada989f1bd","Type":"ContainerDied","Data":"34a5fc34ea87574c9a3b2491be8b0ba8a21c317c01ffc7532a113e83ed1d9145"} Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.835273 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34a5fc34ea87574c9a3b2491be8b0ba8a21c317c01ffc7532a113e83ed1d9145" Nov 24 13:49:27 crc kubenswrapper[4970]: I1124 13:49:27.947376 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j24dc" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="registry-server" probeResult="failure" output=< Nov 24 13:49:27 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 13:49:27 crc kubenswrapper[4970]: > Nov 24 13:49:28 crc kubenswrapper[4970]: I1124 13:49:28.333895 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n8rxn" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="registry-server" probeResult="failure" output=< Nov 24 13:49:28 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 13:49:28 crc kubenswrapper[4970]: > Nov 24 13:49:32 crc kubenswrapper[4970]: I1124 13:49:32.413295 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 13:49:33 crc kubenswrapper[4970]: I1124 13:49:33.924924 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:33 crc kubenswrapper[4970]: I1124 13:49:33.924969 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:33 crc kubenswrapper[4970]: I1124 13:49:33.994246 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:34 crc kubenswrapper[4970]: I1124 13:49:34.121823 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:34 crc kubenswrapper[4970]: I1124 13:49:34.122150 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:34 crc kubenswrapper[4970]: I1124 13:49:34.168358 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:34 crc kubenswrapper[4970]: I1124 13:49:34.906930 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:34 crc kubenswrapper[4970]: I1124 13:49:34.921484 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.292224 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.424223 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.486698 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.487055 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.530915 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.879792 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerStarted","Data":"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94"} Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.926833 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.949003 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:35 crc kubenswrapper[4970]: I1124 13:49:35.997970 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:36 crc kubenswrapper[4970]: I1124 13:49:36.887087 4970 generic.go:334] "Generic (PLEG): container finished" podID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerID="12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94" exitCode=0 Nov 24 13:49:36 crc kubenswrapper[4970]: I1124 13:49:36.888367 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerDied","Data":"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94"} Nov 24 13:49:36 crc kubenswrapper[4970]: I1124 13:49:36.890661 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-49x5r" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="registry-server" containerID="cri-o://1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601" gracePeriod=2 Nov 24 13:49:36 crc kubenswrapper[4970]: I1124 13:49:36.927043 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:49:36 crc kubenswrapper[4970]: I1124 13:49:36.965772 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.223969 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.224854 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c9jhs" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="registry-server" containerID="cri-o://b9412a5c54c6986bc5aab9b182af91c9951d52b65a5ccfa2e45826585709ee20" gracePeriod=2 Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.334055 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:37 crc kubenswrapper[4970]: E1124 13:49:37.378335 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod238eae5b_0891_4035_bd9d_a427bc7d713d.slice/crio-conmon-1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601.scope\": RecentStats: unable to find data in memory cache]" Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.385898 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.893297 4970 generic.go:334] "Generic (PLEG): container finished" podID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerID="b9412a5c54c6986bc5aab9b182af91c9951d52b65a5ccfa2e45826585709ee20" exitCode=0 Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.893381 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerDied","Data":"b9412a5c54c6986bc5aab9b182af91c9951d52b65a5ccfa2e45826585709ee20"} Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.895055 4970 generic.go:334] "Generic (PLEG): container finished" podID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerID="1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601" exitCode=0 Nov 24 13:49:37 crc kubenswrapper[4970]: I1124 13:49:37.895080 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerDied","Data":"1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601"} Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.262224 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.338364 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities\") pod \"238eae5b-0891-4035-bd9d-a427bc7d713d\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.338472 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5tsl\" (UniqueName: \"kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl\") pod \"238eae5b-0891-4035-bd9d-a427bc7d713d\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.338532 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content\") pod \"238eae5b-0891-4035-bd9d-a427bc7d713d\" (UID: \"238eae5b-0891-4035-bd9d-a427bc7d713d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.339062 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities" (OuterVolumeSpecName: "utilities") pod "238eae5b-0891-4035-bd9d-a427bc7d713d" (UID: "238eae5b-0891-4035-bd9d-a427bc7d713d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.358138 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl" (OuterVolumeSpecName: "kube-api-access-j5tsl") pod "238eae5b-0891-4035-bd9d-a427bc7d713d" (UID: "238eae5b-0891-4035-bd9d-a427bc7d713d"). InnerVolumeSpecName "kube-api-access-j5tsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.392928 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "238eae5b-0891-4035-bd9d-a427bc7d713d" (UID: "238eae5b-0891-4035-bd9d-a427bc7d713d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.395267 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.439603 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.439647 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/238eae5b-0891-4035-bd9d-a427bc7d713d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.439661 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5tsl\" (UniqueName: \"kubernetes.io/projected/238eae5b-0891-4035-bd9d-a427bc7d713d-kube-api-access-j5tsl\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.540341 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities\") pod \"973d682e-c2a5-48ed-9787-7f38271e3a6d\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.540404 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blwkq\" (UniqueName: \"kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq\") pod \"973d682e-c2a5-48ed-9787-7f38271e3a6d\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.540460 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content\") pod \"973d682e-c2a5-48ed-9787-7f38271e3a6d\" (UID: \"973d682e-c2a5-48ed-9787-7f38271e3a6d\") " Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.542386 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities" (OuterVolumeSpecName: "utilities") pod "973d682e-c2a5-48ed-9787-7f38271e3a6d" (UID: "973d682e-c2a5-48ed-9787-7f38271e3a6d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.545528 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq" (OuterVolumeSpecName: "kube-api-access-blwkq") pod "973d682e-c2a5-48ed-9787-7f38271e3a6d" (UID: "973d682e-c2a5-48ed-9787-7f38271e3a6d"). InnerVolumeSpecName "kube-api-access-blwkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.600043 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "973d682e-c2a5-48ed-9787-7f38271e3a6d" (UID: "973d682e-c2a5-48ed-9787-7f38271e3a6d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.641990 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.642034 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blwkq\" (UniqueName: \"kubernetes.io/projected/973d682e-c2a5-48ed-9787-7f38271e3a6d-kube-api-access-blwkq\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.642049 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/973d682e-c2a5-48ed-9787-7f38271e3a6d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.912618 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerStarted","Data":"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7"} Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.914953 4970 generic.go:334] "Generic (PLEG): container finished" podID="72e5f1c1-078d-4394-9f04-5541716da154" containerID="e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9" exitCode=0 Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.915009 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerDied","Data":"e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9"} Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.922057 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9jhs" event={"ID":"973d682e-c2a5-48ed-9787-7f38271e3a6d","Type":"ContainerDied","Data":"c74fe18dadb36a317788c15f2c33095402700d6ff3c246ac268747bbb9e69ade"} Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.922112 4970 scope.go:117] "RemoveContainer" containerID="b9412a5c54c6986bc5aab9b182af91c9951d52b65a5ccfa2e45826585709ee20" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.922171 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9jhs" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.928792 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-49x5r" event={"ID":"238eae5b-0891-4035-bd9d-a427bc7d713d","Type":"ContainerDied","Data":"7c46bf27ecfd5c669757457452fd9097d4cdf62511c27843ac3e84589ac0dabf"} Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.928894 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-49x5r" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.938064 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9b6jr" podStartSLOduration=2.165574105 podStartE2EDuration="45.938045477s" podCreationTimestamp="2025-11-24 13:48:53 +0000 UTC" firstStartedPulling="2025-11-24 13:48:54.375182119 +0000 UTC m=+149.662939412" lastFinishedPulling="2025-11-24 13:49:38.147653491 +0000 UTC m=+193.435410784" observedRunningTime="2025-11-24 13:49:38.936276662 +0000 UTC m=+194.224033955" watchObservedRunningTime="2025-11-24 13:49:38.938045477 +0000 UTC m=+194.225802770" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.956120 4970 scope.go:117] "RemoveContainer" containerID="d8f11710cf7a94da9a68e32f05016c6cde6ba68501cb61445b5fd1c4a1e8823f" Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.982486 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.990358 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c9jhs"] Nov 24 13:49:38 crc kubenswrapper[4970]: I1124 13:49:38.996191 4970 scope.go:117] "RemoveContainer" containerID="8fb20d4cbdb416c26d27ce5110a9f701a5d8db0ee607ce464fbc310cf46db546" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.004202 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.006953 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-49x5r"] Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.011997 4970 scope.go:117] "RemoveContainer" containerID="1619f2df720e06b2fa8788f61bd407f7201f6cc1c86c1d7e2ac1528ba8a39601" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.033443 4970 scope.go:117] "RemoveContainer" containerID="71a6230f920181f5efcf73dd4fe3a007c576ef61dac5ffb91d0c1f26477c17db" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.048948 4970 scope.go:117] "RemoveContainer" containerID="dd2235cb2a09b7f1055b9db2fd6319bcbbdcac29f582b6ab9c9311134ec6f096" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.477240 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" path="/var/lib/kubelet/pods/238eae5b-0891-4035-bd9d-a427bc7d713d/volumes" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.477854 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" path="/var/lib/kubelet/pods/973d682e-c2a5-48ed-9787-7f38271e3a6d/volumes" Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.622269 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.622758 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-szhpq" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="registry-server" containerID="cri-o://75578512a5f05ee89528934a829e0627d705ac5d5f361ebe6b1fca298a610f7b" gracePeriod=2 Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.940878 4970 generic.go:334] "Generic (PLEG): container finished" podID="938f9e13-38a8-45db-9e07-1258ced9d931" containerID="75578512a5f05ee89528934a829e0627d705ac5d5f361ebe6b1fca298a610f7b" exitCode=0 Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.940955 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerDied","Data":"75578512a5f05ee89528934a829e0627d705ac5d5f361ebe6b1fca298a610f7b"} Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.952327 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerStarted","Data":"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6"} Nov 24 13:49:39 crc kubenswrapper[4970]: I1124 13:49:39.966501 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h4vlc" podStartSLOduration=1.947918864 podStartE2EDuration="46.966485949s" podCreationTimestamp="2025-11-24 13:48:53 +0000 UTC" firstStartedPulling="2025-11-24 13:48:54.347368495 +0000 UTC m=+149.635125788" lastFinishedPulling="2025-11-24 13:49:39.36593558 +0000 UTC m=+194.653692873" observedRunningTime="2025-11-24 13:49:39.965518893 +0000 UTC m=+195.253276186" watchObservedRunningTime="2025-11-24 13:49:39.966485949 +0000 UTC m=+195.254243242" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.079426 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.157032 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld6jk\" (UniqueName: \"kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk\") pod \"938f9e13-38a8-45db-9e07-1258ced9d931\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.157134 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content\") pod \"938f9e13-38a8-45db-9e07-1258ced9d931\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.157222 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities\") pod \"938f9e13-38a8-45db-9e07-1258ced9d931\" (UID: \"938f9e13-38a8-45db-9e07-1258ced9d931\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.158109 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities" (OuterVolumeSpecName: "utilities") pod "938f9e13-38a8-45db-9e07-1258ced9d931" (UID: "938f9e13-38a8-45db-9e07-1258ced9d931"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.160682 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk" (OuterVolumeSpecName: "kube-api-access-ld6jk") pod "938f9e13-38a8-45db-9e07-1258ced9d931" (UID: "938f9e13-38a8-45db-9e07-1258ced9d931"). InnerVolumeSpecName "kube-api-access-ld6jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.178094 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "938f9e13-38a8-45db-9e07-1258ced9d931" (UID: "938f9e13-38a8-45db-9e07-1258ced9d931"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.226890 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.227160 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n8rxn" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="registry-server" containerID="cri-o://7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4" gracePeriod=2 Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.258929 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.258967 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/938f9e13-38a8-45db-9e07-1258ced9d931-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.258994 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld6jk\" (UniqueName: \"kubernetes.io/projected/938f9e13-38a8-45db-9e07-1258ced9d931-kube-api-access-ld6jk\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.643402 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.763313 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f89x\" (UniqueName: \"kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x\") pod \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.763407 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities\") pod \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.763500 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") pod \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.764763 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities" (OuterVolumeSpecName: "utilities") pod "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" (UID: "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.780130 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x" (OuterVolumeSpecName: "kube-api-access-2f89x") pod "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" (UID: "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e"). InnerVolumeSpecName "kube-api-access-2f89x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.863991 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" (UID: "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.864494 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") pod \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\" (UID: \"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e\") " Nov 24 13:49:40 crc kubenswrapper[4970]: W1124 13:49:40.864721 4970 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e/volumes/kubernetes.io~empty-dir/catalog-content Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.864740 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" (UID: "e3abc2b3-c6fd-41cd-904c-cdbff1044c5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.864851 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f89x\" (UniqueName: \"kubernetes.io/projected/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-kube-api-access-2f89x\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.864877 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.864890 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.958774 4970 generic.go:334] "Generic (PLEG): container finished" podID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerID="7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4" exitCode=0 Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.958845 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerDied","Data":"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4"} Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.958851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8rxn" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.958873 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8rxn" event={"ID":"e3abc2b3-c6fd-41cd-904c-cdbff1044c5e","Type":"ContainerDied","Data":"88a1784032d8ed89f97ade895d103e3fb5ee4665ffcc890fb74760c45acc5cb1"} Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.958888 4970 scope.go:117] "RemoveContainer" containerID="7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.961747 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szhpq" event={"ID":"938f9e13-38a8-45db-9e07-1258ced9d931","Type":"ContainerDied","Data":"80daffd856772ce6a7708f4479d95062c210a0e1b452fd6d878f83a7fe7092ae"} Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.961813 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szhpq" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.980625 4970 scope.go:117] "RemoveContainer" containerID="262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f" Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.984851 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:49:40 crc kubenswrapper[4970]: I1124 13:49:40.986316 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n8rxn"] Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.000316 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.003714 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-szhpq"] Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.010145 4970 scope.go:117] "RemoveContainer" containerID="6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.024225 4970 scope.go:117] "RemoveContainer" containerID="7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4" Nov 24 13:49:41 crc kubenswrapper[4970]: E1124 13:49:41.024639 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4\": container with ID starting with 7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4 not found: ID does not exist" containerID="7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.024692 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4"} err="failed to get container status \"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4\": rpc error: code = NotFound desc = could not find container \"7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4\": container with ID starting with 7eafc3868bebeb28be412a7fdf5aeea86d24540da30cfb1a7e43ddbf73fcaab4 not found: ID does not exist" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.024757 4970 scope.go:117] "RemoveContainer" containerID="262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f" Nov 24 13:49:41 crc kubenswrapper[4970]: E1124 13:49:41.025157 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f\": container with ID starting with 262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f not found: ID does not exist" containerID="262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.025185 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f"} err="failed to get container status \"262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f\": rpc error: code = NotFound desc = could not find container \"262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f\": container with ID starting with 262e85ad8ff4c1a2ef4eb33b19cb9fa99b48ea89b21d1ff471df70219d6f112f not found: ID does not exist" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.025201 4970 scope.go:117] "RemoveContainer" containerID="6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8" Nov 24 13:49:41 crc kubenswrapper[4970]: E1124 13:49:41.025463 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8\": container with ID starting with 6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8 not found: ID does not exist" containerID="6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.025484 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8"} err="failed to get container status \"6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8\": rpc error: code = NotFound desc = could not find container \"6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8\": container with ID starting with 6057e76c949c7117826645c4d5b7a1f54ea8ca6a7fe53edf650ddb19129f47b8 not found: ID does not exist" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.025499 4970 scope.go:117] "RemoveContainer" containerID="75578512a5f05ee89528934a829e0627d705ac5d5f361ebe6b1fca298a610f7b" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.038210 4970 scope.go:117] "RemoveContainer" containerID="93f84f149b321e4246f57f45504f5d37f9768788aa54506f5d8b40f560c99da7" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.052411 4970 scope.go:117] "RemoveContainer" containerID="da7934b9d7ce4b10dec3c85b43e641a751e62c6f0f7b8829b65ae3d3fa9dbec5" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.204283 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.204362 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.477822 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" path="/var/lib/kubelet/pods/938f9e13-38a8-45db-9e07-1258ced9d931/volumes" Nov 24 13:49:41 crc kubenswrapper[4970]: I1124 13:49:41.478700 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" path="/var/lib/kubelet/pods/e3abc2b3-c6fd-41cd-904c-cdbff1044c5e/volumes" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.513006 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.513707 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.561428 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.688247 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.688288 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:49:43 crc kubenswrapper[4970]: I1124 13:49:43.736388 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:49:44 crc kubenswrapper[4970]: I1124 13:49:44.019789 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:49:53 crc kubenswrapper[4970]: I1124 13:49:53.725159 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.337872 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" containerName="oauth-openshift" containerID="cri-o://6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878" gracePeriod=15 Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.825716 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.868986 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-zdnbh"] Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869213 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d809518e-0ed6-43ee-b375-8c63acd0584e" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869228 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d809518e-0ed6-43ee-b375-8c63acd0584e" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869239 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869248 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869259 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" containerName="oauth-openshift" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869267 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" containerName="oauth-openshift" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869279 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869288 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869300 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869308 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869319 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d9dee2-2474-41c7-bb47-dcada989f1bd" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869327 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d9dee2-2474-41c7-bb47-dcada989f1bd" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869337 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869346 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869356 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869365 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869377 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869385 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869394 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869403 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869411 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869418 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="extract-content" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869432 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869440 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869451 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869460 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="extract-utilities" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869473 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869481 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: E1124 13:50:00.869496 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869505 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869648 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d9dee2-2474-41c7-bb47-dcada989f1bd" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869668 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3abc2b3-c6fd-41cd-904c-cdbff1044c5e" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869681 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="238eae5b-0891-4035-bd9d-a427bc7d713d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869696 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d809518e-0ed6-43ee-b375-8c63acd0584e" containerName="pruner" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869714 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" containerName="oauth-openshift" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869726 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="938f9e13-38a8-45db-9e07-1258ced9d931" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.869738 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="973d682e-c2a5-48ed-9787-7f38271e3a6d" containerName="registry-server" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.870197 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.880996 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-zdnbh"] Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.935530 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2vjw\" (UniqueName: \"kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.935863 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.935964 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936020 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936072 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936110 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936144 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936181 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936217 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936245 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936305 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936344 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936361 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936466 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936511 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template\") pod \"60d6d622-9caf-4124-9226-1aed670ffb03\" (UID: \"60d6d622-9caf-4124-9226-1aed670ffb03\") " Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936719 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936741 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936770 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gtp9\" (UniqueName: \"kubernetes.io/projected/aaa751b0-35aa-484c-ad3f-329959a702ac-kube-api-access-9gtp9\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936809 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936840 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936870 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936908 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936946 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936961 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936978 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.936994 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937022 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937018 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937042 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-policies\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937125 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-dir\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937155 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937326 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937523 4970 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937550 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937565 4970 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/60d6d622-9caf-4124-9226-1aed670ffb03-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.937899 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.941771 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.942016 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.942398 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.943298 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.943403 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.943559 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.943667 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw" (OuterVolumeSpecName: "kube-api-access-z2vjw") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "kube-api-access-z2vjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.944062 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:00 crc kubenswrapper[4970]: I1124 13:50:00.944059 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "60d6d622-9caf-4124-9226-1aed670ffb03" (UID: "60d6d622-9caf-4124-9226-1aed670ffb03"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039208 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039276 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039299 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039321 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gtp9\" (UniqueName: \"kubernetes.io/projected/aaa751b0-35aa-484c-ad3f-329959a702ac-kube-api-access-9gtp9\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039345 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039366 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039386 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039411 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039437 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039464 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039479 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039503 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039531 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-policies\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039553 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-dir\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039627 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039645 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039661 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039674 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039687 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039699 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039713 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039724 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039738 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039751 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2vjw\" (UniqueName: \"kubernetes.io/projected/60d6d622-9caf-4124-9226-1aed670ffb03-kube-api-access-z2vjw\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039763 4970 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/60d6d622-9caf-4124-9226-1aed670ffb03-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.039805 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-dir\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.040594 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.040640 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-audit-policies\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.040905 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.041618 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.043489 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.044544 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.046055 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.046234 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.047685 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.047726 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.048643 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.048655 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/aaa751b0-35aa-484c-ad3f-329959a702ac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.068477 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gtp9\" (UniqueName: \"kubernetes.io/projected/aaa751b0-35aa-484c-ad3f-329959a702ac-kube-api-access-9gtp9\") pod \"oauth-openshift-7d9549f6c-zdnbh\" (UID: \"aaa751b0-35aa-484c-ad3f-329959a702ac\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.089057 4970 generic.go:334] "Generic (PLEG): container finished" podID="60d6d622-9caf-4124-9226-1aed670ffb03" containerID="6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878" exitCode=0 Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.089100 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" event={"ID":"60d6d622-9caf-4124-9226-1aed670ffb03","Type":"ContainerDied","Data":"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878"} Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.089131 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" event={"ID":"60d6d622-9caf-4124-9226-1aed670ffb03","Type":"ContainerDied","Data":"0d64c2a055e7f7bf5940ad934145d81b5ca51b1d41fb62999bbdfc83c3e9d531"} Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.089149 4970 scope.go:117] "RemoveContainer" containerID="6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.089547 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lr6pb" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.106737 4970 scope.go:117] "RemoveContainer" containerID="6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878" Nov 24 13:50:01 crc kubenswrapper[4970]: E1124 13:50:01.107435 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878\": container with ID starting with 6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878 not found: ID does not exist" containerID="6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.107563 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878"} err="failed to get container status \"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878\": rpc error: code = NotFound desc = could not find container \"6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878\": container with ID starting with 6b3f84304156ac2324cd9a27499d08ffbf6e84ddc29296c4404467a62c096878 not found: ID does not exist" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.118481 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.120912 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lr6pb"] Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.195860 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.485021 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d6d622-9caf-4124-9226-1aed670ffb03" path="/var/lib/kubelet/pods/60d6d622-9caf-4124-9226-1aed670ffb03/volumes" Nov 24 13:50:01 crc kubenswrapper[4970]: I1124 13:50:01.618462 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-zdnbh"] Nov 24 13:50:02 crc kubenswrapper[4970]: I1124 13:50:02.095791 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" event={"ID":"aaa751b0-35aa-484c-ad3f-329959a702ac","Type":"ContainerStarted","Data":"13a1b82a2ce7c8d21b89e474e742ba8fb4dbc9f3adf48897cae2c6498d107a92"} Nov 24 13:50:02 crc kubenswrapper[4970]: I1124 13:50:02.096103 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:02 crc kubenswrapper[4970]: I1124 13:50:02.096114 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" event={"ID":"aaa751b0-35aa-484c-ad3f-329959a702ac","Type":"ContainerStarted","Data":"f2a7a23b0341db961aaad1a7d48e68886048dc8de648ec893463eb93f62f5077"} Nov 24 13:50:02 crc kubenswrapper[4970]: I1124 13:50:02.113337 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" podStartSLOduration=27.11331705 podStartE2EDuration="27.11331705s" podCreationTimestamp="2025-11-24 13:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:50:02.112321185 +0000 UTC m=+217.400078468" watchObservedRunningTime="2025-11-24 13:50:02.11331705 +0000 UTC m=+217.401074343" Nov 24 13:50:02 crc kubenswrapper[4970]: I1124 13:50:02.369892 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d9549f6c-zdnbh" Nov 24 13:50:11 crc kubenswrapper[4970]: I1124 13:50:11.204565 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:50:11 crc kubenswrapper[4970]: I1124 13:50:11.205162 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:50:11 crc kubenswrapper[4970]: I1124 13:50:11.205210 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:50:11 crc kubenswrapper[4970]: I1124 13:50:11.205798 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:50:11 crc kubenswrapper[4970]: I1124 13:50:11.205861 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833" gracePeriod=600 Nov 24 13:50:12 crc kubenswrapper[4970]: I1124 13:50:12.159703 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833" exitCode=0 Nov 24 13:50:12 crc kubenswrapper[4970]: I1124 13:50:12.159816 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833"} Nov 24 13:50:12 crc kubenswrapper[4970]: I1124 13:50:12.160048 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9"} Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.094048 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.095028 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h4vlc" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="registry-server" containerID="cri-o://56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6" gracePeriod=30 Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.112641 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.112902 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9b6jr" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="registry-server" containerID="cri-o://e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7" gracePeriod=30 Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.129398 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.129702 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" containerID="cri-o://a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73" gracePeriod=30 Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.139551 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.139854 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9pgnr" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="registry-server" containerID="cri-o://85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4" gracePeriod=30 Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.143107 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8njc2"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.144171 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.154381 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.154442 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8njc2"] Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.154633 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j24dc" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="registry-server" containerID="cri-o://5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94" gracePeriod=30 Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.269340 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkgs\" (UniqueName: \"kubernetes.io/projected/9776cf3e-bec6-4eca-874b-37e8dc5a755e-kube-api-access-rjkgs\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.269902 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.269943 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.372353 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.372432 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.372511 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkgs\" (UniqueName: \"kubernetes.io/projected/9776cf3e-bec6-4eca-874b-37e8dc5a755e-kube-api-access-rjkgs\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.374295 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.379521 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9776cf3e-bec6-4eca-874b-37e8dc5a755e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.389844 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkgs\" (UniqueName: \"kubernetes.io/projected/9776cf3e-bec6-4eca-874b-37e8dc5a755e-kube-api-access-rjkgs\") pod \"marketplace-operator-79b997595-8njc2\" (UID: \"9776cf3e-bec6-4eca-874b-37e8dc5a755e\") " pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.521492 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.538699 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.558916 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.603767 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.620851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.637514 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.678681 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktl45\" (UniqueName: \"kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45\") pod \"f026ffe7-2203-429b-9144-e96b21b38a7f\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.679033 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content\") pod \"f026ffe7-2203-429b-9144-e96b21b38a7f\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.679083 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content\") pod \"84c15935-c1f3-4faf-b2a6-a122603f15c7\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.684729 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45" (OuterVolumeSpecName: "kube-api-access-ktl45") pod "f026ffe7-2203-429b-9144-e96b21b38a7f" (UID: "f026ffe7-2203-429b-9144-e96b21b38a7f"). InnerVolumeSpecName "kube-api-access-ktl45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694723 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities\") pod \"f026ffe7-2203-429b-9144-e96b21b38a7f\" (UID: \"f026ffe7-2203-429b-9144-e96b21b38a7f\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694795 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities\") pod \"72e5f1c1-078d-4394-9f04-5541716da154\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694816 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") pod \"8b04d05d-1031-4197-8fb3-59c52252bad7\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694838 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk56x\" (UniqueName: \"kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x\") pod \"8b04d05d-1031-4197-8fb3-59c52252bad7\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694858 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content\") pod \"72e5f1c1-078d-4394-9f04-5541716da154\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694891 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities\") pod \"84c15935-c1f3-4faf-b2a6-a122603f15c7\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694934 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpxhl\" (UniqueName: \"kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl\") pod \"72e5f1c1-078d-4394-9f04-5541716da154\" (UID: \"72e5f1c1-078d-4394-9f04-5541716da154\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694958 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") pod \"8b04d05d-1031-4197-8fb3-59c52252bad7\" (UID: \"8b04d05d-1031-4197-8fb3-59c52252bad7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.694977 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tbxs8\" (UniqueName: \"kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8\") pod \"84c15935-c1f3-4faf-b2a6-a122603f15c7\" (UID: \"84c15935-c1f3-4faf-b2a6-a122603f15c7\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.695299 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktl45\" (UniqueName: \"kubernetes.io/projected/f026ffe7-2203-429b-9144-e96b21b38a7f-kube-api-access-ktl45\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.696262 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities" (OuterVolumeSpecName: "utilities") pod "84c15935-c1f3-4faf-b2a6-a122603f15c7" (UID: "84c15935-c1f3-4faf-b2a6-a122603f15c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.696689 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities" (OuterVolumeSpecName: "utilities") pod "72e5f1c1-078d-4394-9f04-5541716da154" (UID: "72e5f1c1-078d-4394-9f04-5541716da154"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.697681 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8b04d05d-1031-4197-8fb3-59c52252bad7" (UID: "8b04d05d-1031-4197-8fb3-59c52252bad7"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.699110 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities" (OuterVolumeSpecName: "utilities") pod "f026ffe7-2203-429b-9144-e96b21b38a7f" (UID: "f026ffe7-2203-429b-9144-e96b21b38a7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.699622 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8" (OuterVolumeSpecName: "kube-api-access-tbxs8") pod "84c15935-c1f3-4faf-b2a6-a122603f15c7" (UID: "84c15935-c1f3-4faf-b2a6-a122603f15c7"). InnerVolumeSpecName "kube-api-access-tbxs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.700803 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8b04d05d-1031-4197-8fb3-59c52252bad7" (UID: "8b04d05d-1031-4197-8fb3-59c52252bad7"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.701132 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl" (OuterVolumeSpecName: "kube-api-access-zpxhl") pod "72e5f1c1-078d-4394-9f04-5541716da154" (UID: "72e5f1c1-078d-4394-9f04-5541716da154"). InnerVolumeSpecName "kube-api-access-zpxhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.701221 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x" (OuterVolumeSpecName: "kube-api-access-vk56x") pod "8b04d05d-1031-4197-8fb3-59c52252bad7" (UID: "8b04d05d-1031-4197-8fb3-59c52252bad7"). InnerVolumeSpecName "kube-api-access-vk56x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.751928 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f026ffe7-2203-429b-9144-e96b21b38a7f" (UID: "f026ffe7-2203-429b-9144-e96b21b38a7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.785856 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72e5f1c1-078d-4394-9f04-5541716da154" (UID: "72e5f1c1-078d-4394-9f04-5541716da154"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.795797 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwslw\" (UniqueName: \"kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw\") pod \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.795838 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content\") pod \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.795883 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities\") pod \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\" (UID: \"e0fe399e-973f-4c4c-8e49-3619c8788ed2\") " Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796119 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796135 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796145 4970 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796155 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk56x\" (UniqueName: \"kubernetes.io/projected/8b04d05d-1031-4197-8fb3-59c52252bad7-kube-api-access-vk56x\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796163 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72e5f1c1-078d-4394-9f04-5541716da154-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796171 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796178 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpxhl\" (UniqueName: \"kubernetes.io/projected/72e5f1c1-078d-4394-9f04-5541716da154-kube-api-access-zpxhl\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796186 4970 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8b04d05d-1031-4197-8fb3-59c52252bad7-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796196 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tbxs8\" (UniqueName: \"kubernetes.io/projected/84c15935-c1f3-4faf-b2a6-a122603f15c7-kube-api-access-tbxs8\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796204 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f026ffe7-2203-429b-9144-e96b21b38a7f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.796947 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities" (OuterVolumeSpecName: "utilities") pod "e0fe399e-973f-4c4c-8e49-3619c8788ed2" (UID: "e0fe399e-973f-4c4c-8e49-3619c8788ed2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.798408 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw" (OuterVolumeSpecName: "kube-api-access-wwslw") pod "e0fe399e-973f-4c4c-8e49-3619c8788ed2" (UID: "e0fe399e-973f-4c4c-8e49-3619c8788ed2"). InnerVolumeSpecName "kube-api-access-wwslw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.840247 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0fe399e-973f-4c4c-8e49-3619c8788ed2" (UID: "e0fe399e-973f-4c4c-8e49-3619c8788ed2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.852626 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84c15935-c1f3-4faf-b2a6-a122603f15c7" (UID: "84c15935-c1f3-4faf-b2a6-a122603f15c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.896851 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwslw\" (UniqueName: \"kubernetes.io/projected/e0fe399e-973f-4c4c-8e49-3619c8788ed2-kube-api-access-wwslw\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.896885 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.896895 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0fe399e-973f-4c4c-8e49-3619c8788ed2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.896904 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c15935-c1f3-4faf-b2a6-a122603f15c7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:50:17 crc kubenswrapper[4970]: I1124 13:50:17.979188 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8njc2"] Nov 24 13:50:17 crc kubenswrapper[4970]: W1124 13:50:17.984199 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9776cf3e_bec6_4eca_874b_37e8dc5a755e.slice/crio-e15cc6aa37e9e8d0874597b0b4233d9a45ad11a2b3656853cd56a35aebf1319b WatchSource:0}: Error finding container e15cc6aa37e9e8d0874597b0b4233d9a45ad11a2b3656853cd56a35aebf1319b: Status 404 returned error can't find the container with id e15cc6aa37e9e8d0874597b0b4233d9a45ad11a2b3656853cd56a35aebf1319b Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.196936 4970 generic.go:334] "Generic (PLEG): container finished" podID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerID="e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7" exitCode=0 Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.197015 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b6jr" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.197037 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerDied","Data":"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.197095 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b6jr" event={"ID":"f026ffe7-2203-429b-9144-e96b21b38a7f","Type":"ContainerDied","Data":"f6cd24614edc0e68ce89e52984475d91b89e7dfc81dbd90ad33ba922a25fc7cb"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.197122 4970 scope.go:117] "RemoveContainer" containerID="e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.198920 4970 generic.go:334] "Generic (PLEG): container finished" podID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerID="5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94" exitCode=0 Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.198958 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerDied","Data":"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.198985 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j24dc" event={"ID":"84c15935-c1f3-4faf-b2a6-a122603f15c7","Type":"ContainerDied","Data":"62ea9a79f63d6ed14c4fb8502442754547d457ecdf189b43368dc223ddfbb94e"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.198985 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j24dc" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.201493 4970 generic.go:334] "Generic (PLEG): container finished" podID="72e5f1c1-078d-4394-9f04-5541716da154" containerID="56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6" exitCode=0 Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.201545 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h4vlc" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.201620 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerDied","Data":"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.201677 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h4vlc" event={"ID":"72e5f1c1-078d-4394-9f04-5541716da154","Type":"ContainerDied","Data":"74814213c8736640ae0e82105afc001e2774ecbd07315ea17b58433c88ecf745"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.203044 4970 generic.go:334] "Generic (PLEG): container finished" podID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerID="a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73" exitCode=0 Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.203119 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" event={"ID":"8b04d05d-1031-4197-8fb3-59c52252bad7","Type":"ContainerDied","Data":"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.203144 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" event={"ID":"8b04d05d-1031-4197-8fb3-59c52252bad7","Type":"ContainerDied","Data":"36e439283937d8a7377b09d5a13485bc6750d1664d7fc4b76b3eb1b2322850f5"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.203193 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-m968g" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.205846 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerID="85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4" exitCode=0 Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.205910 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerDied","Data":"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.205940 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9pgnr" event={"ID":"e0fe399e-973f-4c4c-8e49-3619c8788ed2","Type":"ContainerDied","Data":"58ecc9cc6be157f94df1e49b2735bc04e1d3aacfccd207f191a834325824ab7c"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.206201 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9pgnr" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.211181 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" event={"ID":"9776cf3e-bec6-4eca-874b-37e8dc5a755e","Type":"ContainerStarted","Data":"e5de05bd64d84ceb609ba365147f56b91783c6f19fde13e2539a0eddb858796d"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.211228 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" event={"ID":"9776cf3e-bec6-4eca-874b-37e8dc5a755e","Type":"ContainerStarted","Data":"e15cc6aa37e9e8d0874597b0b4233d9a45ad11a2b3656853cd56a35aebf1319b"} Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.211562 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.215454 4970 scope.go:117] "RemoveContainer" containerID="12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.216741 4970 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8njc2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.216781 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" podUID="9776cf3e-bec6-4eca-874b-37e8dc5a755e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.240013 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" podStartSLOduration=1.239992665 podStartE2EDuration="1.239992665s" podCreationTimestamp="2025-11-24 13:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:50:18.237938183 +0000 UTC m=+233.525695476" watchObservedRunningTime="2025-11-24 13:50:18.239992665 +0000 UTC m=+233.527749948" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.246826 4970 scope.go:117] "RemoveContainer" containerID="bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.259454 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.262912 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-m968g"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.271485 4970 scope.go:117] "RemoveContainer" containerID="e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.272057 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7\": container with ID starting with e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7 not found: ID does not exist" containerID="e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.272101 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7"} err="failed to get container status \"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7\": rpc error: code = NotFound desc = could not find container \"e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7\": container with ID starting with e2d00d791f6f2087c2e5b13c7e9a5c72fcf1db3d8295a2bafe6d12472a789ed7 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.272127 4970 scope.go:117] "RemoveContainer" containerID="12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.272253 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.272447 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94\": container with ID starting with 12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94 not found: ID does not exist" containerID="12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.272478 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94"} err="failed to get container status \"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94\": rpc error: code = NotFound desc = could not find container \"12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94\": container with ID starting with 12fd435657f587e0080983036d0492ea449b48a851c79bfb26e3b44000a07b94 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.272498 4970 scope.go:117] "RemoveContainer" containerID="bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.274894 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee\": container with ID starting with bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee not found: ID does not exist" containerID="bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.274960 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee"} err="failed to get container status \"bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee\": rpc error: code = NotFound desc = could not find container \"bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee\": container with ID starting with bca0efb0d4ae77dbd8b6cebc53860555d90a061c71995f35b93cec02eb0961ee not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.275042 4970 scope.go:117] "RemoveContainer" containerID="5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.275976 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9pgnr"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.290013 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.294398 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h4vlc"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.300214 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.306647 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9b6jr"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.310964 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.316359 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j24dc"] Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.327018 4970 scope.go:117] "RemoveContainer" containerID="c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.339409 4970 scope.go:117] "RemoveContainer" containerID="3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.354558 4970 scope.go:117] "RemoveContainer" containerID="5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.355064 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94\": container with ID starting with 5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94 not found: ID does not exist" containerID="5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.355108 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94"} err="failed to get container status \"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94\": rpc error: code = NotFound desc = could not find container \"5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94\": container with ID starting with 5eaea24509a5c4acb9e07ff616bafecbcd7e1780f1159baff8252539d7ca5c94 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.355154 4970 scope.go:117] "RemoveContainer" containerID="c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.355511 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd\": container with ID starting with c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd not found: ID does not exist" containerID="c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.355549 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd"} err="failed to get container status \"c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd\": rpc error: code = NotFound desc = could not find container \"c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd\": container with ID starting with c842ede880214513b818b02c95e6d79a940591f6f0efb68d3375c27b82492dfd not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.355593 4970 scope.go:117] "RemoveContainer" containerID="3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.355970 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c\": container with ID starting with 3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c not found: ID does not exist" containerID="3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.355994 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c"} err="failed to get container status \"3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c\": rpc error: code = NotFound desc = could not find container \"3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c\": container with ID starting with 3555a65ff949a69916f25d165483f2e8ee1ac275640085015a4d5c82f7536f9c not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.356013 4970 scope.go:117] "RemoveContainer" containerID="56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.372162 4970 scope.go:117] "RemoveContainer" containerID="e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.388855 4970 scope.go:117] "RemoveContainer" containerID="07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.407027 4970 scope.go:117] "RemoveContainer" containerID="56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.408540 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6\": container with ID starting with 56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6 not found: ID does not exist" containerID="56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.408610 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6"} err="failed to get container status \"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6\": rpc error: code = NotFound desc = could not find container \"56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6\": container with ID starting with 56e906d9b33423ca0d2f0010b1ba7615ecad99fab89c7a5905805893b6960aa6 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.408644 4970 scope.go:117] "RemoveContainer" containerID="e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.409070 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9\": container with ID starting with e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9 not found: ID does not exist" containerID="e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.409114 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9"} err="failed to get container status \"e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9\": rpc error: code = NotFound desc = could not find container \"e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9\": container with ID starting with e4057f524a49683445d5add7e8cb0d6967bf1c64f1bef7c1835d0b185a29fea9 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.409143 4970 scope.go:117] "RemoveContainer" containerID="07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.409377 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf\": container with ID starting with 07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf not found: ID does not exist" containerID="07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.409407 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf"} err="failed to get container status \"07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf\": rpc error: code = NotFound desc = could not find container \"07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf\": container with ID starting with 07095fd5a4a8b49f109367c4b4e971aea4636c070dddfc779a22582113c38adf not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.409424 4970 scope.go:117] "RemoveContainer" containerID="a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.420427 4970 scope.go:117] "RemoveContainer" containerID="a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.420729 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73\": container with ID starting with a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73 not found: ID does not exist" containerID="a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.420761 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73"} err="failed to get container status \"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73\": rpc error: code = NotFound desc = could not find container \"a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73\": container with ID starting with a94af75561628528fcc4650f945480d21173ab46ca4ec2a74ef1c44b78a91a73 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.420780 4970 scope.go:117] "RemoveContainer" containerID="85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.431301 4970 scope.go:117] "RemoveContainer" containerID="b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.445060 4970 scope.go:117] "RemoveContainer" containerID="f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.459259 4970 scope.go:117] "RemoveContainer" containerID="85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.459650 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4\": container with ID starting with 85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4 not found: ID does not exist" containerID="85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.459675 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4"} err="failed to get container status \"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4\": rpc error: code = NotFound desc = could not find container \"85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4\": container with ID starting with 85aa044b94470a4968f7a1a2bb8803c1a3313e522497a1f196733659406b76f4 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.459695 4970 scope.go:117] "RemoveContainer" containerID="b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.460428 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552\": container with ID starting with b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552 not found: ID does not exist" containerID="b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.460462 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552"} err="failed to get container status \"b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552\": rpc error: code = NotFound desc = could not find container \"b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552\": container with ID starting with b6e5322cb0e6e3067c49b40fe2cd5942d0af1dad1ab77cc396547c6c226f2552 not found: ID does not exist" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.460476 4970 scope.go:117] "RemoveContainer" containerID="f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77" Nov 24 13:50:18 crc kubenswrapper[4970]: E1124 13:50:18.460750 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77\": container with ID starting with f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77 not found: ID does not exist" containerID="f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77" Nov 24 13:50:18 crc kubenswrapper[4970]: I1124 13:50:18.460770 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77"} err="failed to get container status \"f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77\": rpc error: code = NotFound desc = could not find container \"f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77\": container with ID starting with f3b0f485c3a4eee90d3e0036fd5ba01962223274ebee0d9dc8663671e8e52e77 not found: ID does not exist" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.223829 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8njc2" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315308 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2lldf"] Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315526 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315543 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315554 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315562 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315643 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315655 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315672 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315680 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315692 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315699 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315712 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315719 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315731 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315739 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315748 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315756 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315765 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315771 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="extract-content" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315782 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315790 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315799 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315806 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315814 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315820 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: E1124 13:50:19.315830 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315837 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="extract-utilities" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315964 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" containerName="marketplace-operator" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315980 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.315992 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.316004 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e5f1c1-078d-4394-9f04-5541716da154" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.316015 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" containerName="registry-server" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.317034 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.319355 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.321300 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lldf"] Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.415645 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-utilities\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.415742 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5r9r\" (UniqueName: \"kubernetes.io/projected/17f79817-a5af-4690-8524-716178cec162-kube-api-access-c5r9r\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.415774 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-catalog-content\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.477708 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e5f1c1-078d-4394-9f04-5541716da154" path="/var/lib/kubelet/pods/72e5f1c1-078d-4394-9f04-5541716da154/volumes" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.478676 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84c15935-c1f3-4faf-b2a6-a122603f15c7" path="/var/lib/kubelet/pods/84c15935-c1f3-4faf-b2a6-a122603f15c7/volumes" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.479515 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b04d05d-1031-4197-8fb3-59c52252bad7" path="/var/lib/kubelet/pods/8b04d05d-1031-4197-8fb3-59c52252bad7/volumes" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.480816 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0fe399e-973f-4c4c-8e49-3619c8788ed2" path="/var/lib/kubelet/pods/e0fe399e-973f-4c4c-8e49-3619c8788ed2/volumes" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.481605 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f026ffe7-2203-429b-9144-e96b21b38a7f" path="/var/lib/kubelet/pods/f026ffe7-2203-429b-9144-e96b21b38a7f/volumes" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.512913 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w8gm8"] Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.517254 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.518314 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-utilities\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.520272 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-utilities\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.520631 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.523067 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w8gm8"] Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.527082 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5r9r\" (UniqueName: \"kubernetes.io/projected/17f79817-a5af-4690-8524-716178cec162-kube-api-access-c5r9r\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.527158 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-catalog-content\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.528558 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f79817-a5af-4690-8524-716178cec162-catalog-content\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.599904 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5r9r\" (UniqueName: \"kubernetes.io/projected/17f79817-a5af-4690-8524-716178cec162-kube-api-access-c5r9r\") pod \"redhat-marketplace-2lldf\" (UID: \"17f79817-a5af-4690-8524-716178cec162\") " pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.628269 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-catalog-content\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.628313 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skx7v\" (UniqueName: \"kubernetes.io/projected/745f7b30-991a-4df2-bdae-d7438d628a7f-kube-api-access-skx7v\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.628357 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-utilities\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.633628 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.729666 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-utilities\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.729755 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-catalog-content\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.729813 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skx7v\" (UniqueName: \"kubernetes.io/projected/745f7b30-991a-4df2-bdae-d7438d628a7f-kube-api-access-skx7v\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.730557 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-utilities\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.730690 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/745f7b30-991a-4df2-bdae-d7438d628a7f-catalog-content\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.755279 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skx7v\" (UniqueName: \"kubernetes.io/projected/745f7b30-991a-4df2-bdae-d7438d628a7f-kube-api-access-skx7v\") pod \"certified-operators-w8gm8\" (UID: \"745f7b30-991a-4df2-bdae-d7438d628a7f\") " pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.804429 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lldf"] Nov 24 13:50:19 crc kubenswrapper[4970]: I1124 13:50:19.919498 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.096076 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w8gm8"] Nov 24 13:50:20 crc kubenswrapper[4970]: W1124 13:50:20.100920 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod745f7b30_991a_4df2_bdae_d7438d628a7f.slice/crio-bbfe0af2db415c93ad50c748275a943ab41931708833af7b6b6aa152666a655d WatchSource:0}: Error finding container bbfe0af2db415c93ad50c748275a943ab41931708833af7b6b6aa152666a655d: Status 404 returned error can't find the container with id bbfe0af2db415c93ad50c748275a943ab41931708833af7b6b6aa152666a655d Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.228455 4970 generic.go:334] "Generic (PLEG): container finished" podID="17f79817-a5af-4690-8524-716178cec162" containerID="3f45cb3178015ac0a07a6f258705d2637689e5536ac1a1d5d04dc978ed7f23a2" exitCode=0 Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.228556 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lldf" event={"ID":"17f79817-a5af-4690-8524-716178cec162","Type":"ContainerDied","Data":"3f45cb3178015ac0a07a6f258705d2637689e5536ac1a1d5d04dc978ed7f23a2"} Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.228609 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lldf" event={"ID":"17f79817-a5af-4690-8524-716178cec162","Type":"ContainerStarted","Data":"7320e689f73410be74c313e17c97c3dbd64cfa687bda05f16b5a628099397cf8"} Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.230137 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8gm8" event={"ID":"745f7b30-991a-4df2-bdae-d7438d628a7f","Type":"ContainerStarted","Data":"abcf18fab22f171d40600afad8d97b35ad0db81fe80552d1c48f8d226c25bdfd"} Nov 24 13:50:20 crc kubenswrapper[4970]: I1124 13:50:20.230183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8gm8" event={"ID":"745f7b30-991a-4df2-bdae-d7438d628a7f","Type":"ContainerStarted","Data":"bbfe0af2db415c93ad50c748275a943ab41931708833af7b6b6aa152666a655d"} Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.243743 4970 generic.go:334] "Generic (PLEG): container finished" podID="17f79817-a5af-4690-8524-716178cec162" containerID="c3e1fdabf232c54cffe752304cd2df3bca34d7bb57b32f60b21abc452a678ec3" exitCode=0 Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.243979 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lldf" event={"ID":"17f79817-a5af-4690-8524-716178cec162","Type":"ContainerDied","Data":"c3e1fdabf232c54cffe752304cd2df3bca34d7bb57b32f60b21abc452a678ec3"} Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.251019 4970 generic.go:334] "Generic (PLEG): container finished" podID="745f7b30-991a-4df2-bdae-d7438d628a7f" containerID="abcf18fab22f171d40600afad8d97b35ad0db81fe80552d1c48f8d226c25bdfd" exitCode=0 Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.251069 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8gm8" event={"ID":"745f7b30-991a-4df2-bdae-d7438d628a7f","Type":"ContainerDied","Data":"abcf18fab22f171d40600afad8d97b35ad0db81fe80552d1c48f8d226c25bdfd"} Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.711060 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.712023 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.714776 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.719995 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.754155 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.754191 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.754260 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdvnx\" (UniqueName: \"kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.854941 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdvnx\" (UniqueName: \"kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.854997 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.855021 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.855481 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.856065 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.880523 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdvnx\" (UniqueName: \"kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx\") pod \"redhat-operators-frtlk\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.914913 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.916100 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.922785 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.927366 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.956409 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.956454 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4wzh\" (UniqueName: \"kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:21 crc kubenswrapper[4970]: I1124 13:50:21.956502 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.039650 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.057929 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.057968 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4wzh\" (UniqueName: \"kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.058006 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.058762 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.058797 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.077208 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4wzh\" (UniqueName: \"kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh\") pod \"community-operators-xwtxh\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.251507 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.257831 4970 generic.go:334] "Generic (PLEG): container finished" podID="745f7b30-991a-4df2-bdae-d7438d628a7f" containerID="65c9474d3be252db7ad059f15de92dd17ae72db2768c02c49e11bedd1ff79deb" exitCode=0 Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.257913 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8gm8" event={"ID":"745f7b30-991a-4df2-bdae-d7438d628a7f","Type":"ContainerDied","Data":"65c9474d3be252db7ad059f15de92dd17ae72db2768c02c49e11bedd1ff79deb"} Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.270167 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lldf" event={"ID":"17f79817-a5af-4690-8524-716178cec162","Type":"ContainerStarted","Data":"c02aef74a50b15472a9caa54eba64ce44d9a6bf3a34abb4a5d904289a11e9649"} Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.299345 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2lldf" podStartSLOduration=1.902746636 podStartE2EDuration="3.299311789s" podCreationTimestamp="2025-11-24 13:50:19 +0000 UTC" firstStartedPulling="2025-11-24 13:50:20.231260902 +0000 UTC m=+235.519018195" lastFinishedPulling="2025-11-24 13:50:21.627826055 +0000 UTC m=+236.915583348" observedRunningTime="2025-11-24 13:50:22.297457072 +0000 UTC m=+237.585214365" watchObservedRunningTime="2025-11-24 13:50:22.299311789 +0000 UTC m=+237.587069082" Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.432360 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:50:22 crc kubenswrapper[4970]: W1124 13:50:22.439440 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda73c8c7e_4407_4624_93cc_c13e6d56347a.slice/crio-187f4ecb5c66eadea0066bb269d82b2245a90b60c4b02179b3651571c9caf16f WatchSource:0}: Error finding container 187f4ecb5c66eadea0066bb269d82b2245a90b60c4b02179b3651571c9caf16f: Status 404 returned error can't find the container with id 187f4ecb5c66eadea0066bb269d82b2245a90b60c4b02179b3651571c9caf16f Nov 24 13:50:22 crc kubenswrapper[4970]: I1124 13:50:22.691239 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 13:50:22 crc kubenswrapper[4970]: W1124 13:50:22.695849 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ff82462_7b38_4888_af1f_8d5eb8629c58.slice/crio-5aeaac152f7a43bd795f5c39e073b1c905738e7144a24637962065594ed584bb WatchSource:0}: Error finding container 5aeaac152f7a43bd795f5c39e073b1c905738e7144a24637962065594ed584bb: Status 404 returned error can't find the container with id 5aeaac152f7a43bd795f5c39e073b1c905738e7144a24637962065594ed584bb Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.276290 4970 generic.go:334] "Generic (PLEG): container finished" podID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerID="3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6" exitCode=0 Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.276383 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerDied","Data":"3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6"} Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.276813 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerStarted","Data":"187f4ecb5c66eadea0066bb269d82b2245a90b60c4b02179b3651571c9caf16f"} Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.280255 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w8gm8" event={"ID":"745f7b30-991a-4df2-bdae-d7438d628a7f","Type":"ContainerStarted","Data":"be05a5bdd20c4c336ca2249851ef480d6c4641ce83f068de008dd5961c7ae596"} Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.281481 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerID="6e0214bae0969f87cd8dd43cdc6c88ea6fe79774c1eac911d6e1941016de6aea" exitCode=0 Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.281596 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerDied","Data":"6e0214bae0969f87cd8dd43cdc6c88ea6fe79774c1eac911d6e1941016de6aea"} Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.281629 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerStarted","Data":"5aeaac152f7a43bd795f5c39e073b1c905738e7144a24637962065594ed584bb"} Nov 24 13:50:23 crc kubenswrapper[4970]: I1124 13:50:23.329708 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w8gm8" podStartSLOduration=2.9147992179999997 podStartE2EDuration="4.329691334s" podCreationTimestamp="2025-11-24 13:50:19 +0000 UTC" firstStartedPulling="2025-11-24 13:50:21.252882021 +0000 UTC m=+236.540639314" lastFinishedPulling="2025-11-24 13:50:22.667774137 +0000 UTC m=+237.955531430" observedRunningTime="2025-11-24 13:50:23.328943416 +0000 UTC m=+238.616700719" watchObservedRunningTime="2025-11-24 13:50:23.329691334 +0000 UTC m=+238.617448627" Nov 24 13:50:24 crc kubenswrapper[4970]: I1124 13:50:24.292210 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerStarted","Data":"6689f03422c6a1dd3ea5001899e6af473996e5d53336134f25b16d306b9dee2c"} Nov 24 13:50:25 crc kubenswrapper[4970]: I1124 13:50:25.300532 4970 generic.go:334] "Generic (PLEG): container finished" podID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerID="c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1" exitCode=0 Nov 24 13:50:25 crc kubenswrapper[4970]: I1124 13:50:25.300637 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerDied","Data":"c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1"} Nov 24 13:50:25 crc kubenswrapper[4970]: I1124 13:50:25.304103 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerID="6689f03422c6a1dd3ea5001899e6af473996e5d53336134f25b16d306b9dee2c" exitCode=0 Nov 24 13:50:25 crc kubenswrapper[4970]: I1124 13:50:25.304152 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerDied","Data":"6689f03422c6a1dd3ea5001899e6af473996e5d53336134f25b16d306b9dee2c"} Nov 24 13:50:27 crc kubenswrapper[4970]: I1124 13:50:27.319434 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerStarted","Data":"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0"} Nov 24 13:50:27 crc kubenswrapper[4970]: I1124 13:50:27.326007 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerStarted","Data":"66e8c6276601fd762d7293cc6127dce3f778bc2bb0de3c6104b446c5a16f90c6"} Nov 24 13:50:27 crc kubenswrapper[4970]: I1124 13:50:27.345648 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frtlk" podStartSLOduration=3.763078078 podStartE2EDuration="6.34562273s" podCreationTimestamp="2025-11-24 13:50:21 +0000 UTC" firstStartedPulling="2025-11-24 13:50:23.278044293 +0000 UTC m=+238.565801586" lastFinishedPulling="2025-11-24 13:50:25.860588935 +0000 UTC m=+241.148346238" observedRunningTime="2025-11-24 13:50:27.339943294 +0000 UTC m=+242.627700597" watchObservedRunningTime="2025-11-24 13:50:27.34562273 +0000 UTC m=+242.633380063" Nov 24 13:50:27 crc kubenswrapper[4970]: I1124 13:50:27.357638 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xwtxh" podStartSLOduration=3.665770597 podStartE2EDuration="6.357618009s" podCreationTimestamp="2025-11-24 13:50:21 +0000 UTC" firstStartedPulling="2025-11-24 13:50:23.284231992 +0000 UTC m=+238.571989285" lastFinishedPulling="2025-11-24 13:50:25.976079404 +0000 UTC m=+241.263836697" observedRunningTime="2025-11-24 13:50:27.355145025 +0000 UTC m=+242.642902318" watchObservedRunningTime="2025-11-24 13:50:27.357618009 +0000 UTC m=+242.645375302" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.634432 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.634840 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.693483 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.920356 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.920408 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:29 crc kubenswrapper[4970]: I1124 13:50:29.965433 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:30 crc kubenswrapper[4970]: I1124 13:50:30.391194 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w8gm8" Nov 24 13:50:30 crc kubenswrapper[4970]: I1124 13:50:30.393634 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2lldf" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.040689 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.040750 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.090521 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.252602 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.252988 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.288795 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.383632 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 13:50:32 crc kubenswrapper[4970]: I1124 13:50:32.402467 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:52:11 crc kubenswrapper[4970]: I1124 13:52:11.204525 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:52:11 crc kubenswrapper[4970]: I1124 13:52:11.206058 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:52:41 crc kubenswrapper[4970]: I1124 13:52:41.204231 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:52:41 crc kubenswrapper[4970]: I1124 13:52:41.204895 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.204805 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.205359 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.205418 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.206262 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.206354 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9" gracePeriod=600 Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.384209 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9" exitCode=0 Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.384290 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9"} Nov 24 13:53:11 crc kubenswrapper[4970]: I1124 13:53:11.384503 4970 scope.go:117] "RemoveContainer" containerID="948c9a550039ada255640568668c0fc6de0558fe3e5552e58a357d0a06764833" Nov 24 13:53:12 crc kubenswrapper[4970]: I1124 13:53:12.395510 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272"} Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.494079 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9ml9j"] Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.494875 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.513285 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9ml9j"] Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665029 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk2pv\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-kube-api-access-qk2pv\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665085 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d8bec5d-aa86-49bc-8556-7296d170a0f0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665143 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-trusted-ca\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665169 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d8bec5d-aa86-49bc-8556-7296d170a0f0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665196 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-tls\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665227 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665252 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-certificates\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.665275 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-bound-sa-token\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.686102 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766187 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d8bec5d-aa86-49bc-8556-7296d170a0f0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766273 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-tls\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766330 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-certificates\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766373 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-bound-sa-token\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766418 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk2pv\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-kube-api-access-qk2pv\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766501 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d8bec5d-aa86-49bc-8556-7296d170a0f0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.766560 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-trusted-ca\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.767978 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1d8bec5d-aa86-49bc-8556-7296d170a0f0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.768833 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-trusted-ca\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.768851 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-certificates\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.776685 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1d8bec5d-aa86-49bc-8556-7296d170a0f0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.777169 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-registry-tls\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.783782 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-bound-sa-token\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.786646 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk2pv\" (UniqueName: \"kubernetes.io/projected/1d8bec5d-aa86-49bc-8556-7296d170a0f0-kube-api-access-qk2pv\") pod \"image-registry-66df7c8f76-9ml9j\" (UID: \"1d8bec5d-aa86-49bc-8556-7296d170a0f0\") " pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:14 crc kubenswrapper[4970]: I1124 13:53:14.816834 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:15 crc kubenswrapper[4970]: I1124 13:53:15.044039 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9ml9j"] Nov 24 13:53:15 crc kubenswrapper[4970]: W1124 13:53:15.054925 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d8bec5d_aa86_49bc_8556_7296d170a0f0.slice/crio-b762effa917d459197999f306b42edbe8ed275b6c24007b28ddd03cf24ad7263 WatchSource:0}: Error finding container b762effa917d459197999f306b42edbe8ed275b6c24007b28ddd03cf24ad7263: Status 404 returned error can't find the container with id b762effa917d459197999f306b42edbe8ed275b6c24007b28ddd03cf24ad7263 Nov 24 13:53:15 crc kubenswrapper[4970]: I1124 13:53:15.421392 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" event={"ID":"1d8bec5d-aa86-49bc-8556-7296d170a0f0","Type":"ContainerStarted","Data":"e6fb55be963e18f5f5a1ccc0a751d04cc257da873c51573d672a27fccf203f51"} Nov 24 13:53:15 crc kubenswrapper[4970]: I1124 13:53:15.421849 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" event={"ID":"1d8bec5d-aa86-49bc-8556-7296d170a0f0","Type":"ContainerStarted","Data":"b762effa917d459197999f306b42edbe8ed275b6c24007b28ddd03cf24ad7263"} Nov 24 13:53:15 crc kubenswrapper[4970]: I1124 13:53:15.422783 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:15 crc kubenswrapper[4970]: I1124 13:53:15.458339 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" podStartSLOduration=1.4583156179999999 podStartE2EDuration="1.458315618s" podCreationTimestamp="2025-11-24 13:53:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:53:15.455406324 +0000 UTC m=+410.743163617" watchObservedRunningTime="2025-11-24 13:53:15.458315618 +0000 UTC m=+410.746072931" Nov 24 13:53:34 crc kubenswrapper[4970]: I1124 13:53:34.825771 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9ml9j" Nov 24 13:53:34 crc kubenswrapper[4970]: I1124 13:53:34.902482 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:53:59 crc kubenswrapper[4970]: I1124 13:53:59.955851 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" podUID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" containerName="registry" containerID="cri-o://2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613" gracePeriod=30 Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.371492 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544392 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544457 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544513 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544557 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544625 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544672 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7fmb\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544708 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.544746 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets\") pod \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\" (UID: \"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8\") " Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.545569 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.545952 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.551476 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.551832 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb" (OuterVolumeSpecName: "kube-api-access-x7fmb") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "kube-api-access-x7fmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.558757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.559386 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.561417 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.566866 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" (UID: "7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646377 4970 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646423 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646437 4970 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646451 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7fmb\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-kube-api-access-x7fmb\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646462 4970 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646475 4970 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.646486 4970 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.720332 4970 generic.go:334] "Generic (PLEG): container finished" podID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" containerID="2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613" exitCode=0 Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.720381 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" event={"ID":"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8","Type":"ContainerDied","Data":"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613"} Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.720454 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" event={"ID":"7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8","Type":"ContainerDied","Data":"8287ef6534faf9eb3d568fcb4f2e9b94d37ee1cb04d3a0e048f7f46989d46889"} Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.720466 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rgct9" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.720479 4970 scope.go:117] "RemoveContainer" containerID="2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.737690 4970 scope.go:117] "RemoveContainer" containerID="2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613" Nov 24 13:54:00 crc kubenswrapper[4970]: E1124 13:54:00.738256 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613\": container with ID starting with 2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613 not found: ID does not exist" containerID="2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.738307 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613"} err="failed to get container status \"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613\": rpc error: code = NotFound desc = could not find container \"2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613\": container with ID starting with 2040692ca2cd59a139c34ab872607e9e7d302e55133c51bd2c371a3c421a3613 not found: ID does not exist" Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.768708 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:54:00 crc kubenswrapper[4970]: I1124 13:54:00.774087 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rgct9"] Nov 24 13:54:01 crc kubenswrapper[4970]: I1124 13:54:01.480748 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" path="/var/lib/kubelet/pods/7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8/volumes" Nov 24 13:55:11 crc kubenswrapper[4970]: I1124 13:55:11.204127 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:55:11 crc kubenswrapper[4970]: I1124 13:55:11.204915 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.979970 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-22n4h"] Nov 24 13:55:32 crc kubenswrapper[4970]: E1124 13:55:32.980624 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" containerName="registry" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.980635 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" containerName="registry" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.980739 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5d2b6d-30d7-4504-8fb8-b11e7eb576d8" containerName="registry" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.981071 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.984687 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.987800 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.987940 4970 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-mg9g5" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.992083 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bz6q5"] Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.992790 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.994918 4970 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-ct2x5" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.995948 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cjd48"] Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.996731 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-cjd48" Nov 24 13:55:32 crc kubenswrapper[4970]: I1124 13:55:32.998834 4970 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-ndqrj" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.010132 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bz6q5"] Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.027066 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cjd48"] Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.029994 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-22n4h"] Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.144817 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjkh7\" (UniqueName: \"kubernetes.io/projected/5da63529-7e53-4ad3-8d4b-0bc32d616af8-kube-api-access-qjkh7\") pod \"cert-manager-cainjector-7f985d654d-22n4h\" (UID: \"5da63529-7e53-4ad3-8d4b-0bc32d616af8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.144902 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8txz5\" (UniqueName: \"kubernetes.io/projected/fe636fdb-0614-4cbc-a621-086b3c591926-kube-api-access-8txz5\") pod \"cert-manager-webhook-5655c58dd6-bz6q5\" (UID: \"fe636fdb-0614-4cbc-a621-086b3c591926\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.144920 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdwxn\" (UniqueName: \"kubernetes.io/projected/45373789-c5a3-47eb-b2fe-7b4ce5236a33-kube-api-access-gdwxn\") pod \"cert-manager-5b446d88c5-cjd48\" (UID: \"45373789-c5a3-47eb-b2fe-7b4ce5236a33\") " pod="cert-manager/cert-manager-5b446d88c5-cjd48" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.246518 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8txz5\" (UniqueName: \"kubernetes.io/projected/fe636fdb-0614-4cbc-a621-086b3c591926-kube-api-access-8txz5\") pod \"cert-manager-webhook-5655c58dd6-bz6q5\" (UID: \"fe636fdb-0614-4cbc-a621-086b3c591926\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.246564 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdwxn\" (UniqueName: \"kubernetes.io/projected/45373789-c5a3-47eb-b2fe-7b4ce5236a33-kube-api-access-gdwxn\") pod \"cert-manager-5b446d88c5-cjd48\" (UID: \"45373789-c5a3-47eb-b2fe-7b4ce5236a33\") " pod="cert-manager/cert-manager-5b446d88c5-cjd48" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.246644 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjkh7\" (UniqueName: \"kubernetes.io/projected/5da63529-7e53-4ad3-8d4b-0bc32d616af8-kube-api-access-qjkh7\") pod \"cert-manager-cainjector-7f985d654d-22n4h\" (UID: \"5da63529-7e53-4ad3-8d4b-0bc32d616af8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.265499 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8txz5\" (UniqueName: \"kubernetes.io/projected/fe636fdb-0614-4cbc-a621-086b3c591926-kube-api-access-8txz5\") pod \"cert-manager-webhook-5655c58dd6-bz6q5\" (UID: \"fe636fdb-0614-4cbc-a621-086b3c591926\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.267219 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjkh7\" (UniqueName: \"kubernetes.io/projected/5da63529-7e53-4ad3-8d4b-0bc32d616af8-kube-api-access-qjkh7\") pod \"cert-manager-cainjector-7f985d654d-22n4h\" (UID: \"5da63529-7e53-4ad3-8d4b-0bc32d616af8\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.274179 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdwxn\" (UniqueName: \"kubernetes.io/projected/45373789-c5a3-47eb-b2fe-7b4ce5236a33-kube-api-access-gdwxn\") pod \"cert-manager-5b446d88c5-cjd48\" (UID: \"45373789-c5a3-47eb-b2fe-7b4ce5236a33\") " pod="cert-manager/cert-manager-5b446d88c5-cjd48" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.297987 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.324953 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.328004 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-cjd48" Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.594463 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bz6q5"] Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.605176 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.624252 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cjd48"] Nov 24 13:55:33 crc kubenswrapper[4970]: W1124 13:55:33.628246 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45373789_c5a3_47eb_b2fe_7b4ce5236a33.slice/crio-a2ff42b04cb6436d76386baadc95aaa12e9331eb047adbc56f06186c60dafb16 WatchSource:0}: Error finding container a2ff42b04cb6436d76386baadc95aaa12e9331eb047adbc56f06186c60dafb16: Status 404 returned error can't find the container with id a2ff42b04cb6436d76386baadc95aaa12e9331eb047adbc56f06186c60dafb16 Nov 24 13:55:33 crc kubenswrapper[4970]: I1124 13:55:33.744465 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-22n4h"] Nov 24 13:55:33 crc kubenswrapper[4970]: W1124 13:55:33.750881 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5da63529_7e53_4ad3_8d4b_0bc32d616af8.slice/crio-de0a45904574af6a289a1ee1a6be5b2239c9dcd49777cadd0ac4865abeaf7032 WatchSource:0}: Error finding container de0a45904574af6a289a1ee1a6be5b2239c9dcd49777cadd0ac4865abeaf7032: Status 404 returned error can't find the container with id de0a45904574af6a289a1ee1a6be5b2239c9dcd49777cadd0ac4865abeaf7032 Nov 24 13:55:34 crc kubenswrapper[4970]: I1124 13:55:34.403692 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" event={"ID":"fe636fdb-0614-4cbc-a621-086b3c591926","Type":"ContainerStarted","Data":"95b3cdb8157f59f46ce0ecea78bb7645fb3aa072300463193363832695e24611"} Nov 24 13:55:34 crc kubenswrapper[4970]: I1124 13:55:34.404657 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-cjd48" event={"ID":"45373789-c5a3-47eb-b2fe-7b4ce5236a33","Type":"ContainerStarted","Data":"a2ff42b04cb6436d76386baadc95aaa12e9331eb047adbc56f06186c60dafb16"} Nov 24 13:55:34 crc kubenswrapper[4970]: I1124 13:55:34.405449 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" event={"ID":"5da63529-7e53-4ad3-8d4b-0bc32d616af8","Type":"ContainerStarted","Data":"de0a45904574af6a289a1ee1a6be5b2239c9dcd49777cadd0ac4865abeaf7032"} Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.435712 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" event={"ID":"5da63529-7e53-4ad3-8d4b-0bc32d616af8","Type":"ContainerStarted","Data":"b1efa1629f20f14fa7b21aa48887a8f517511370ae3b8ae5b06b969859c40c01"} Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.440277 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" event={"ID":"fe636fdb-0614-4cbc-a621-086b3c591926","Type":"ContainerStarted","Data":"3eaa46f894503f1f137709415c75346cad6bfb39112525a7a47638dca67ba4cc"} Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.440435 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.444300 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-cjd48" event={"ID":"45373789-c5a3-47eb-b2fe-7b4ce5236a33","Type":"ContainerStarted","Data":"f7724b163e180f318a936bdbf374f6b48b30ae895b55954061a5efa1a873d664"} Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.451838 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-22n4h" podStartSLOduration=2.256061667 podStartE2EDuration="5.451821051s" podCreationTimestamp="2025-11-24 13:55:32 +0000 UTC" firstStartedPulling="2025-11-24 13:55:33.753793305 +0000 UTC m=+549.041550598" lastFinishedPulling="2025-11-24 13:55:36.949552679 +0000 UTC m=+552.237309982" observedRunningTime="2025-11-24 13:55:37.450013555 +0000 UTC m=+552.737770848" watchObservedRunningTime="2025-11-24 13:55:37.451821051 +0000 UTC m=+552.739578344" Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.465633 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" podStartSLOduration=2.118261983 podStartE2EDuration="5.465610253s" podCreationTimestamp="2025-11-24 13:55:32 +0000 UTC" firstStartedPulling="2025-11-24 13:55:33.604808206 +0000 UTC m=+548.892565499" lastFinishedPulling="2025-11-24 13:55:36.952156476 +0000 UTC m=+552.239913769" observedRunningTime="2025-11-24 13:55:37.46467827 +0000 UTC m=+552.752435563" watchObservedRunningTime="2025-11-24 13:55:37.465610253 +0000 UTC m=+552.753367556" Nov 24 13:55:37 crc kubenswrapper[4970]: I1124 13:55:37.481192 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-cjd48" podStartSLOduration=2.151980277 podStartE2EDuration="5.481175801s" podCreationTimestamp="2025-11-24 13:55:32 +0000 UTC" firstStartedPulling="2025-11-24 13:55:33.630913764 +0000 UTC m=+548.918671067" lastFinishedPulling="2025-11-24 13:55:36.960109298 +0000 UTC m=+552.247866591" observedRunningTime="2025-11-24 13:55:37.479999542 +0000 UTC m=+552.767756835" watchObservedRunningTime="2025-11-24 13:55:37.481175801 +0000 UTC m=+552.768933084" Nov 24 13:55:41 crc kubenswrapper[4970]: I1124 13:55:41.204332 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:55:41 crc kubenswrapper[4970]: I1124 13:55:41.204667 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304084 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-msd5k"] Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304608 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-controller" containerID="cri-o://37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304660 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="northd" containerID="cri-o://f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304739 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-acl-logging" containerID="cri-o://a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304684 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="sbdb" containerID="cri-o://aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304777 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-node" containerID="cri-o://54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304797 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="nbdb" containerID="cri-o://39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.304817 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.332528 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-bz6q5" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.360907 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" containerID="cri-o://0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" gracePeriod=30 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.485802 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/2.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.489623 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/1.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.489670 4970 generic.go:334] "Generic (PLEG): container finished" podID="0875a754-321a-4a3a-a798-941eefcdffe5" containerID="3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde" exitCode=2 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.489743 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerDied","Data":"3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde"} Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.489790 4970 scope.go:117] "RemoveContainer" containerID="53a2dbec1ce1b023cec72be5e9c67c4906f76edd00b8bc495d47db1301e6151a" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.490221 4970 scope.go:117] "RemoveContainer" containerID="3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.490395 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vtfnd_openshift-multus(0875a754-321a-4a3a-a798-941eefcdffe5)\"" pod="openshift-multus/multus-vtfnd" podUID="0875a754-321a-4a3a-a798-941eefcdffe5" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.494416 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.496363 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-acl-logging/0.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.496737 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-controller/0.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497032 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" exitCode=0 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497054 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" exitCode=0 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497063 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" exitCode=143 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497072 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" exitCode=143 Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497091 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497115 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497124 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.497132 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.637420 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.640454 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-acl-logging/0.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.641233 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-controller/0.log" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.641594 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707412 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s7h9b"] Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707633 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-node" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707645 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-node" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707653 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kubecfg-setup" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707660 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kubecfg-setup" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707667 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707672 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707682 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707689 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707697 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-acl-logging" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707703 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-acl-logging" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707710 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="sbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707716 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="sbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707722 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="northd" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707727 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="northd" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707737 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="nbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707743 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="nbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707754 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707760 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707770 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707776 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707782 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707788 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.707795 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707801 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707886 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707895 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="sbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707904 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="northd" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707912 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707918 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="nbdb" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707926 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-acl-logging" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707934 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707941 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovn-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707949 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-node" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.707956 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 13:55:43 crc kubenswrapper[4970]: E1124 13:55:43.708037 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.708044 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.708126 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.708136 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerName="ovnkube-controller" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.709549 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.810816 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811070 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811153 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.810962 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811204 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811221 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811259 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811321 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811373 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811394 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811411 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811437 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811464 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811488 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811437 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash" (OuterVolumeSpecName: "host-slash") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811531 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811566 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811621 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811647 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811687 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r754h\" (UniqueName: \"kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811633 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811633 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log" (OuterVolumeSpecName: "node-log") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811654 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811681 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811716 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811761 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811778 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811821 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811846 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811872 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib\") pod \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\" (UID: \"7b38f92e-6c99-47b9-9079-0b0e455ddca1\") " Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811875 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket" (OuterVolumeSpecName: "log-socket") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811908 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811934 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.811950 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812090 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812169 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812272 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812341 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-node-log\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812304 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812456 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-script-lib\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812557 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812664 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-slash\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812744 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812812 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-ovn\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812878 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-bin\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.812989 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc8gx\" (UniqueName: \"kubernetes.io/projected/3dd45d90-4717-4488-b458-6a89c1582dce-kube-api-access-fc8gx\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813037 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-netd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813096 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-config\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813162 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3dd45d90-4717-4488-b458-6a89c1582dce-ovn-node-metrics-cert\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813264 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-kubelet\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813331 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-netns\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813394 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-env-overrides\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813493 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-systemd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813606 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-var-lib-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813691 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-systemd-units\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813759 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-etc-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813830 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-log-socket\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.813919 4970 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814002 4970 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814067 4970 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814134 4970 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814190 4970 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814242 4970 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814298 4970 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814352 4970 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814479 4970 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814533 4970 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814616 4970 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814693 4970 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814756 4970 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814814 4970 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7b38f92e-6c99-47b9-9079-0b0e455ddca1-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814867 4970 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814922 4970 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.814973 4970 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.818375 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.818906 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h" (OuterVolumeSpecName: "kube-api-access-r754h") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "kube-api-access-r754h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.834724 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "7b38f92e-6c99-47b9-9079-0b0e455ddca1" (UID: "7b38f92e-6c99-47b9-9079-0b0e455ddca1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.915897 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.915976 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-slash\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916014 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916047 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916053 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-ovn\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916106 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-ovn\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916114 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-bin\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916162 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-ovn-kubernetes\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916158 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-slash\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916216 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-netd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916138 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-bin\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916179 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-cni-netd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916302 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc8gx\" (UniqueName: \"kubernetes.io/projected/3dd45d90-4717-4488-b458-6a89c1582dce-kube-api-access-fc8gx\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916446 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-config\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916518 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3dd45d90-4717-4488-b458-6a89c1582dce-ovn-node-metrics-cert\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916570 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-kubelet\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916632 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-netns\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916665 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-env-overrides\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916710 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-var-lib-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916753 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-kubelet\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916756 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-systemd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916813 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-systemd\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916837 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-etc-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916876 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-systemd-units\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916885 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-var-lib-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916941 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-log-socket\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916946 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-etc-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.916979 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-systemd-units\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917010 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-host-run-netns\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917037 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917074 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-node-log\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917079 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-run-openvswitch\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917106 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-script-lib\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917142 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-node-log\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917095 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3dd45d90-4717-4488-b458-6a89c1582dce-log-socket\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917336 4970 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b38f92e-6c99-47b9-9079-0b0e455ddca1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917377 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r754h\" (UniqueName: \"kubernetes.io/projected/7b38f92e-6c99-47b9-9079-0b0e455ddca1-kube-api-access-r754h\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917410 4970 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7b38f92e-6c99-47b9-9079-0b0e455ddca1-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.917558 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-env-overrides\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.918094 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-config\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.918388 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3dd45d90-4717-4488-b458-6a89c1582dce-ovnkube-script-lib\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.923447 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3dd45d90-4717-4488-b458-6a89c1582dce-ovn-node-metrics-cert\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:43 crc kubenswrapper[4970]: I1124 13:55:43.947898 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc8gx\" (UniqueName: \"kubernetes.io/projected/3dd45d90-4717-4488-b458-6a89c1582dce-kube-api-access-fc8gx\") pod \"ovnkube-node-s7h9b\" (UID: \"3dd45d90-4717-4488-b458-6a89c1582dce\") " pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.029561 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.505616 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/2.log" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.509212 4970 generic.go:334] "Generic (PLEG): container finished" podID="3dd45d90-4717-4488-b458-6a89c1582dce" containerID="e3a346029c6d197c021042129dceda6ecaa5afb02b3303b6eae0965785391dc5" exitCode=0 Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.509313 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerDied","Data":"e3a346029c6d197c021042129dceda6ecaa5afb02b3303b6eae0965785391dc5"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.509361 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"cf63dfdcc8fa6b0e9183f729ec34464dbb5215b24dad5069d264d9d4b241b6b1"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.515014 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovnkube-controller/3.log" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.520901 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-acl-logging/0.log" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.521751 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-msd5k_7b38f92e-6c99-47b9-9079-0b0e455ddca1/ovn-controller/0.log" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.522560 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" exitCode=0 Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.522778 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" exitCode=0 Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.522904 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" exitCode=0 Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523022 4970 generic.go:334] "Generic (PLEG): container finished" podID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" exitCode=0 Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523173 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523356 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523508 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523758 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.524012 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" event={"ID":"7b38f92e-6c99-47b9-9079-0b0e455ddca1","Type":"ContainerDied","Data":"9ca80d3c309203ce36121caca75393fc8f197217c0a4bf60d0cb60c84fb45dc3"} Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523923 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-msd5k" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.523885 4970 scope.go:117] "RemoveContainer" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.559908 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.581715 4970 scope.go:117] "RemoveContainer" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.631289 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-msd5k"] Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.631910 4970 scope.go:117] "RemoveContainer" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.649506 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-msd5k"] Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.661777 4970 scope.go:117] "RemoveContainer" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.689065 4970 scope.go:117] "RemoveContainer" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.709861 4970 scope.go:117] "RemoveContainer" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.736689 4970 scope.go:117] "RemoveContainer" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.767938 4970 scope.go:117] "RemoveContainer" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.791517 4970 scope.go:117] "RemoveContainer" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.827916 4970 scope.go:117] "RemoveContainer" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.828458 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": container with ID starting with 0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055 not found: ID does not exist" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.828510 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} err="failed to get container status \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": rpc error: code = NotFound desc = could not find container \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": container with ID starting with 0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.828538 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.829317 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": container with ID starting with 2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36 not found: ID does not exist" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.829358 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} err="failed to get container status \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": rpc error: code = NotFound desc = could not find container \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": container with ID starting with 2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.829389 4970 scope.go:117] "RemoveContainer" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.829809 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": container with ID starting with aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575 not found: ID does not exist" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.829871 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} err="failed to get container status \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": rpc error: code = NotFound desc = could not find container \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": container with ID starting with aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.829913 4970 scope.go:117] "RemoveContainer" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.830761 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": container with ID starting with 39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0 not found: ID does not exist" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.830804 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} err="failed to get container status \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": rpc error: code = NotFound desc = could not find container \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": container with ID starting with 39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.830829 4970 scope.go:117] "RemoveContainer" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.831205 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": container with ID starting with f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda not found: ID does not exist" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.831239 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} err="failed to get container status \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": rpc error: code = NotFound desc = could not find container \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": container with ID starting with f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.831263 4970 scope.go:117] "RemoveContainer" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.831735 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": container with ID starting with f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486 not found: ID does not exist" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.831767 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} err="failed to get container status \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": rpc error: code = NotFound desc = could not find container \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": container with ID starting with f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.831791 4970 scope.go:117] "RemoveContainer" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.832197 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": container with ID starting with 54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463 not found: ID does not exist" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.832247 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} err="failed to get container status \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": rpc error: code = NotFound desc = could not find container \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": container with ID starting with 54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.832276 4970 scope.go:117] "RemoveContainer" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.832916 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": container with ID starting with a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476 not found: ID does not exist" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.832952 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} err="failed to get container status \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": rpc error: code = NotFound desc = could not find container \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": container with ID starting with a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.832970 4970 scope.go:117] "RemoveContainer" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.833405 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": container with ID starting with 37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983 not found: ID does not exist" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.833450 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} err="failed to get container status \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": rpc error: code = NotFound desc = could not find container \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": container with ID starting with 37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.833480 4970 scope.go:117] "RemoveContainer" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: E1124 13:55:44.834265 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": container with ID starting with a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed not found: ID does not exist" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.834299 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed"} err="failed to get container status \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": rpc error: code = NotFound desc = could not find container \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": container with ID starting with a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.834320 4970 scope.go:117] "RemoveContainer" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.835485 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} err="failed to get container status \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": rpc error: code = NotFound desc = could not find container \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": container with ID starting with 0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.835515 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.835940 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} err="failed to get container status \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": rpc error: code = NotFound desc = could not find container \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": container with ID starting with 2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.835986 4970 scope.go:117] "RemoveContainer" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.836661 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} err="failed to get container status \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": rpc error: code = NotFound desc = could not find container \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": container with ID starting with aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.836692 4970 scope.go:117] "RemoveContainer" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.837280 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} err="failed to get container status \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": rpc error: code = NotFound desc = could not find container \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": container with ID starting with 39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.837307 4970 scope.go:117] "RemoveContainer" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.837837 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} err="failed to get container status \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": rpc error: code = NotFound desc = could not find container \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": container with ID starting with f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.837860 4970 scope.go:117] "RemoveContainer" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.838227 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} err="failed to get container status \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": rpc error: code = NotFound desc = could not find container \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": container with ID starting with f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.838280 4970 scope.go:117] "RemoveContainer" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.838771 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} err="failed to get container status \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": rpc error: code = NotFound desc = could not find container \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": container with ID starting with 54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.838810 4970 scope.go:117] "RemoveContainer" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839178 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} err="failed to get container status \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": rpc error: code = NotFound desc = could not find container \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": container with ID starting with a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839271 4970 scope.go:117] "RemoveContainer" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839563 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} err="failed to get container status \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": rpc error: code = NotFound desc = could not find container \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": container with ID starting with 37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839612 4970 scope.go:117] "RemoveContainer" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839890 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed"} err="failed to get container status \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": rpc error: code = NotFound desc = could not find container \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": container with ID starting with a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.839927 4970 scope.go:117] "RemoveContainer" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.840213 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} err="failed to get container status \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": rpc error: code = NotFound desc = could not find container \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": container with ID starting with 0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.840249 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841079 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} err="failed to get container status \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": rpc error: code = NotFound desc = could not find container \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": container with ID starting with 2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841104 4970 scope.go:117] "RemoveContainer" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841484 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} err="failed to get container status \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": rpc error: code = NotFound desc = could not find container \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": container with ID starting with aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841513 4970 scope.go:117] "RemoveContainer" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841874 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} err="failed to get container status \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": rpc error: code = NotFound desc = could not find container \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": container with ID starting with 39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.841909 4970 scope.go:117] "RemoveContainer" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.842305 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} err="failed to get container status \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": rpc error: code = NotFound desc = could not find container \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": container with ID starting with f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.842345 4970 scope.go:117] "RemoveContainer" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.842720 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} err="failed to get container status \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": rpc error: code = NotFound desc = could not find container \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": container with ID starting with f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.842761 4970 scope.go:117] "RemoveContainer" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843084 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} err="failed to get container status \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": rpc error: code = NotFound desc = could not find container \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": container with ID starting with 54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843113 4970 scope.go:117] "RemoveContainer" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843495 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} err="failed to get container status \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": rpc error: code = NotFound desc = could not find container \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": container with ID starting with a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843527 4970 scope.go:117] "RemoveContainer" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843910 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} err="failed to get container status \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": rpc error: code = NotFound desc = could not find container \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": container with ID starting with 37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.843952 4970 scope.go:117] "RemoveContainer" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.844346 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed"} err="failed to get container status \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": rpc error: code = NotFound desc = could not find container \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": container with ID starting with a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.844380 4970 scope.go:117] "RemoveContainer" containerID="0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.844827 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055"} err="failed to get container status \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": rpc error: code = NotFound desc = could not find container \"0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055\": container with ID starting with 0e2b5821807ac3fe0de05b8016ce1246cdef32544fe5ca1b17026cce19bc5055 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.844866 4970 scope.go:117] "RemoveContainer" containerID="2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845181 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36"} err="failed to get container status \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": rpc error: code = NotFound desc = could not find container \"2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36\": container with ID starting with 2afcb3609fb172839e2f4ac0b53063ebee14b1c49410f33f7d99c05352bd6c36 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845223 4970 scope.go:117] "RemoveContainer" containerID="aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845495 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575"} err="failed to get container status \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": rpc error: code = NotFound desc = could not find container \"aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575\": container with ID starting with aa2bf0f0f4cc22be161ede753cb46d268f9c6854443afc1a7623503344600575 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845531 4970 scope.go:117] "RemoveContainer" containerID="39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845931 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0"} err="failed to get container status \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": rpc error: code = NotFound desc = could not find container \"39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0\": container with ID starting with 39111427d5461524cc7cf1328594fcdbdc5e29af85c4b6ed0261011a7cf463e0 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.845970 4970 scope.go:117] "RemoveContainer" containerID="f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.846286 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda"} err="failed to get container status \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": rpc error: code = NotFound desc = could not find container \"f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda\": container with ID starting with f4c8f8a2a00a8c9a9c874f3c81865f3a4fd2a55da67225233975151d97d62fda not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.846348 4970 scope.go:117] "RemoveContainer" containerID="f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.846698 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486"} err="failed to get container status \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": rpc error: code = NotFound desc = could not find container \"f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486\": container with ID starting with f2a3adaac2e0ed1646cc4b35dedf6a57b9a1561737f9d573402ca79a67cb1486 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.846728 4970 scope.go:117] "RemoveContainer" containerID="54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.846985 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463"} err="failed to get container status \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": rpc error: code = NotFound desc = could not find container \"54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463\": container with ID starting with 54a9f0a985b0c08dc2d197bb5ab8aecd640309cb6757df678a500ff4a9cea463 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.847020 4970 scope.go:117] "RemoveContainer" containerID="a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.847292 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476"} err="failed to get container status \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": rpc error: code = NotFound desc = could not find container \"a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476\": container with ID starting with a958a44d3860cd60bda2e5edffe138877a7ddac70e4c5d2ac307db0d2437e476 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.847329 4970 scope.go:117] "RemoveContainer" containerID="37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.847753 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983"} err="failed to get container status \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": rpc error: code = NotFound desc = could not find container \"37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983\": container with ID starting with 37b2b83ec0e5141a84ea0632020da979ed5fc2e88beae96786ded9b2e7086983 not found: ID does not exist" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.847791 4970 scope.go:117] "RemoveContainer" containerID="a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed" Nov 24 13:55:44 crc kubenswrapper[4970]: I1124 13:55:44.848138 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed"} err="failed to get container status \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": rpc error: code = NotFound desc = could not find container \"a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed\": container with ID starting with a4ba7f3b2c42ff830f05d27ccd0b8a883e8e03512eb52fb09b9ca5fac939d0ed not found: ID does not exist" Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.478097 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b38f92e-6c99-47b9-9079-0b0e455ddca1" path="/var/lib/kubelet/pods/7b38f92e-6c99-47b9-9079-0b0e455ddca1/volumes" Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534680 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"22c2cede354761f311e51faea5831f401e984d90fb64e575c65768f5038c1cd2"} Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534765 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"391e97177dd5e7b93b79bef02f62f24510ad746df9fc744c56ab5caa1c0cb80a"} Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534793 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"a5436a8be1244ce26533dbc3dbf35d4f2dc26bca58829e6f0c2ca1ed98ef6237"} Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534817 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"c9a9c15fb3e387921292eedc09b2f1231e74b4c8d49a6d54b0448c762e362db2"} Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534839 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"6cfa056afd7e8caa9a515b7a808b2df4ff596e33c4e2bbd0c41ea9aa101eaf86"} Nov 24 13:55:45 crc kubenswrapper[4970]: I1124 13:55:45.534867 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"91895334d55d956ad77e8885136ebfccb90d0e1c5191fc309c995f538a6fb18c"} Nov 24 13:55:48 crc kubenswrapper[4970]: I1124 13:55:48.556038 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"6d8148c504c996beead533c1d11920c66970a771c5e5bc62c4bab8cbb7b6c7cf"} Nov 24 13:55:50 crc kubenswrapper[4970]: I1124 13:55:50.572514 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" event={"ID":"3dd45d90-4717-4488-b458-6a89c1582dce","Type":"ContainerStarted","Data":"f4443ed171769ef894b7e1d3bc34d54c5173aedb1fc83f33bc580a92bfede580"} Nov 24 13:55:50 crc kubenswrapper[4970]: I1124 13:55:50.573060 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:50 crc kubenswrapper[4970]: I1124 13:55:50.573072 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:50 crc kubenswrapper[4970]: I1124 13:55:50.615434 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:50 crc kubenswrapper[4970]: I1124 13:55:50.622492 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" podStartSLOduration=7.622475354 podStartE2EDuration="7.622475354s" podCreationTimestamp="2025-11-24 13:55:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:55:50.619277623 +0000 UTC m=+565.907034926" watchObservedRunningTime="2025-11-24 13:55:50.622475354 +0000 UTC m=+565.910232647" Nov 24 13:55:51 crc kubenswrapper[4970]: I1124 13:55:51.580003 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:51 crc kubenswrapper[4970]: I1124 13:55:51.620969 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:55:55 crc kubenswrapper[4970]: I1124 13:55:55.476177 4970 scope.go:117] "RemoveContainer" containerID="3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde" Nov 24 13:55:55 crc kubenswrapper[4970]: E1124 13:55:55.476962 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-vtfnd_openshift-multus(0875a754-321a-4a3a-a798-941eefcdffe5)\"" pod="openshift-multus/multus-vtfnd" podUID="0875a754-321a-4a3a-a798-941eefcdffe5" Nov 24 13:56:06 crc kubenswrapper[4970]: I1124 13:56:06.469875 4970 scope.go:117] "RemoveContainer" containerID="3d9658d08186ef213c0fdaf19bd276416748af4bd86591008c650f12ee50cfde" Nov 24 13:56:07 crc kubenswrapper[4970]: I1124 13:56:07.693401 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-vtfnd_0875a754-321a-4a3a-a798-941eefcdffe5/kube-multus/2.log" Nov 24 13:56:07 crc kubenswrapper[4970]: I1124 13:56:07.693816 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-vtfnd" event={"ID":"0875a754-321a-4a3a-a798-941eefcdffe5","Type":"ContainerStarted","Data":"112a721b17f6648b444ddaafa4637d86aa71b6be36f3c1948dcab9e6ab74b56a"} Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.204419 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.205234 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.205308 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.206188 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.206367 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272" gracePeriod=600 Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.737810 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272" exitCode=0 Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.737858 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272"} Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.738552 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099"} Nov 24 13:56:11 crc kubenswrapper[4970]: I1124 13:56:11.738627 4970 scope.go:117] "RemoveContainer" containerID="d0ae09240f3430ae45452f82fb38f9f7c815e8107d2641225e2098fb265215c9" Nov 24 13:56:14 crc kubenswrapper[4970]: I1124 13:56:14.060645 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s7h9b" Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.817761 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx"] Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.820445 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.823503 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.828317 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx"] Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.961173 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjjrj\" (UniqueName: \"kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.961227 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:22 crc kubenswrapper[4970]: I1124 13:56:22.961283 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.062734 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.062863 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.062887 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjjrj\" (UniqueName: \"kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.063808 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.063882 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.099446 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjjrj\" (UniqueName: \"kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.148153 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.386444 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx"] Nov 24 13:56:23 crc kubenswrapper[4970]: W1124 13:56:23.396707 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda21ef8b8_372b_45fb_93e3_d0318862eab0.slice/crio-5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774 WatchSource:0}: Error finding container 5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774: Status 404 returned error can't find the container with id 5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774 Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.827367 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerStarted","Data":"bdfa528b39a048fa3640f4132affae53ab422b87fdf485f88dbc432c992dd950"} Nov 24 13:56:23 crc kubenswrapper[4970]: I1124 13:56:23.827418 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerStarted","Data":"5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774"} Nov 24 13:56:24 crc kubenswrapper[4970]: I1124 13:56:24.837312 4970 generic.go:334] "Generic (PLEG): container finished" podID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerID="bdfa528b39a048fa3640f4132affae53ab422b87fdf485f88dbc432c992dd950" exitCode=0 Nov 24 13:56:24 crc kubenswrapper[4970]: I1124 13:56:24.838992 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerDied","Data":"bdfa528b39a048fa3640f4132affae53ab422b87fdf485f88dbc432c992dd950"} Nov 24 13:56:26 crc kubenswrapper[4970]: I1124 13:56:26.853471 4970 generic.go:334] "Generic (PLEG): container finished" podID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerID="cb0f1b0bd55293c73629d56a6a1c6138640eee2c9532e0b99ccf8ba1dcc5f274" exitCode=0 Nov 24 13:56:26 crc kubenswrapper[4970]: I1124 13:56:26.853527 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerDied","Data":"cb0f1b0bd55293c73629d56a6a1c6138640eee2c9532e0b99ccf8ba1dcc5f274"} Nov 24 13:56:27 crc kubenswrapper[4970]: I1124 13:56:27.865304 4970 generic.go:334] "Generic (PLEG): container finished" podID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerID="f7074bdb3f65943bf3a04831ac547bb9c40106487d5b42eebbd75b0c9e435212" exitCode=0 Nov 24 13:56:27 crc kubenswrapper[4970]: I1124 13:56:27.865386 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerDied","Data":"f7074bdb3f65943bf3a04831ac547bb9c40106487d5b42eebbd75b0c9e435212"} Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.165712 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.349280 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjjrj\" (UniqueName: \"kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj\") pod \"a21ef8b8-372b-45fb-93e3-d0318862eab0\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.349412 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util\") pod \"a21ef8b8-372b-45fb-93e3-d0318862eab0\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.349459 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle\") pod \"a21ef8b8-372b-45fb-93e3-d0318862eab0\" (UID: \"a21ef8b8-372b-45fb-93e3-d0318862eab0\") " Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.350253 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle" (OuterVolumeSpecName: "bundle") pod "a21ef8b8-372b-45fb-93e3-d0318862eab0" (UID: "a21ef8b8-372b-45fb-93e3-d0318862eab0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.358860 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj" (OuterVolumeSpecName: "kube-api-access-fjjrj") pod "a21ef8b8-372b-45fb-93e3-d0318862eab0" (UID: "a21ef8b8-372b-45fb-93e3-d0318862eab0"). InnerVolumeSpecName "kube-api-access-fjjrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.434309 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util" (OuterVolumeSpecName: "util") pod "a21ef8b8-372b-45fb-93e3-d0318862eab0" (UID: "a21ef8b8-372b-45fb-93e3-d0318862eab0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.450707 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjjrj\" (UniqueName: \"kubernetes.io/projected/a21ef8b8-372b-45fb-93e3-d0318862eab0-kube-api-access-fjjrj\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.450746 4970 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.450757 4970 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a21ef8b8-372b-45fb-93e3-d0318862eab0-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.881934 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" event={"ID":"a21ef8b8-372b-45fb-93e3-d0318862eab0","Type":"ContainerDied","Data":"5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774"} Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.882229 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5526efd8c53c9284381c4707e8125e6e06d369e89e376e914d7419c749f67774" Nov 24 13:56:29 crc kubenswrapper[4970]: I1124 13:56:29.882028 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.392790 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-fdjzs"] Nov 24 13:56:34 crc kubenswrapper[4970]: E1124 13:56:34.393261 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="extract" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.393272 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="extract" Nov 24 13:56:34 crc kubenswrapper[4970]: E1124 13:56:34.393288 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="util" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.393294 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="util" Nov 24 13:56:34 crc kubenswrapper[4970]: E1124 13:56:34.393305 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="pull" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.393311 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="pull" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.393412 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a21ef8b8-372b-45fb-93e3-d0318862eab0" containerName="extract" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.393761 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.395607 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-f9hl8" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.398349 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.398706 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.416689 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-fdjzs"] Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.521864 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2mc7\" (UniqueName: \"kubernetes.io/projected/63629bdf-0eb6-46d8-a10b-08cc587486a1-kube-api-access-q2mc7\") pod \"nmstate-operator-557fdffb88-fdjzs\" (UID: \"63629bdf-0eb6-46d8-a10b-08cc587486a1\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.623094 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2mc7\" (UniqueName: \"kubernetes.io/projected/63629bdf-0eb6-46d8-a10b-08cc587486a1-kube-api-access-q2mc7\") pod \"nmstate-operator-557fdffb88-fdjzs\" (UID: \"63629bdf-0eb6-46d8-a10b-08cc587486a1\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.643083 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2mc7\" (UniqueName: \"kubernetes.io/projected/63629bdf-0eb6-46d8-a10b-08cc587486a1-kube-api-access-q2mc7\") pod \"nmstate-operator-557fdffb88-fdjzs\" (UID: \"63629bdf-0eb6-46d8-a10b-08cc587486a1\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.708409 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" Nov 24 13:56:34 crc kubenswrapper[4970]: I1124 13:56:34.957496 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-fdjzs"] Nov 24 13:56:35 crc kubenswrapper[4970]: I1124 13:56:35.915785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" event={"ID":"63629bdf-0eb6-46d8-a10b-08cc587486a1","Type":"ContainerStarted","Data":"005e86b6fa4d93851745f4f38cc56ffbddb8d2674def57fa99c78fbda61032bb"} Nov 24 13:56:37 crc kubenswrapper[4970]: I1124 13:56:37.928148 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" event={"ID":"63629bdf-0eb6-46d8-a10b-08cc587486a1","Type":"ContainerStarted","Data":"e79bebea071ede4107e384f651e629097c59437cf70df06b1b38646df226a74b"} Nov 24 13:56:37 crc kubenswrapper[4970]: I1124 13:56:37.944612 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-fdjzs" podStartSLOduration=1.8599862900000002 podStartE2EDuration="3.944593886s" podCreationTimestamp="2025-11-24 13:56:34 +0000 UTC" firstStartedPulling="2025-11-24 13:56:34.968679464 +0000 UTC m=+610.256436767" lastFinishedPulling="2025-11-24 13:56:37.05328707 +0000 UTC m=+612.341044363" observedRunningTime="2025-11-24 13:56:37.941994293 +0000 UTC m=+613.229751586" watchObservedRunningTime="2025-11-24 13:56:37.944593886 +0000 UTC m=+613.232351179" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.233242 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.234452 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.243863 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-swm29" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.252050 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.258595 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.259648 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.261270 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.286303 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.296035 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fpsdv"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.296963 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.340137 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6c6k\" (UniqueName: \"kubernetes.io/projected/62a1e00f-7e1a-4e97-a311-5628f78a7798-kube-api-access-m6c6k\") pod \"nmstate-metrics-5dcf9c57c5-8nmq4\" (UID: \"62a1e00f-7e1a-4e97-a311-5628f78a7798\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.391253 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.392216 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.394606 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-f6qmv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.395940 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.395947 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.405968 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441201 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441266 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-ovs-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441292 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-dbus-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441355 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-nmstate-lock\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441396 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zm9g\" (UniqueName: \"kubernetes.io/projected/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-kube-api-access-5zm9g\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441439 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6c6k\" (UniqueName: \"kubernetes.io/projected/62a1e00f-7e1a-4e97-a311-5628f78a7798-kube-api-access-m6c6k\") pod \"nmstate-metrics-5dcf9c57c5-8nmq4\" (UID: \"62a1e00f-7e1a-4e97-a311-5628f78a7798\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.441497 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g28h5\" (UniqueName: \"kubernetes.io/projected/892ed300-f0bb-49aa-80e3-b7feb492606c-kube-api-access-g28h5\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.472714 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6c6k\" (UniqueName: \"kubernetes.io/projected/62a1e00f-7e1a-4e97-a311-5628f78a7798-kube-api-access-m6c6k\") pod \"nmstate-metrics-5dcf9c57c5-8nmq4\" (UID: \"62a1e00f-7e1a-4e97-a311-5628f78a7798\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542555 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g28h5\" (UniqueName: \"kubernetes.io/projected/892ed300-f0bb-49aa-80e3-b7feb492606c-kube-api-access-g28h5\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542671 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv6dr\" (UniqueName: \"kubernetes.io/projected/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-kube-api-access-xv6dr\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542710 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542756 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-ovs-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542789 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-dbus-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542821 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542844 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-nmstate-lock\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: E1124 13:56:43.542852 4970 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542867 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zm9g\" (UniqueName: \"kubernetes.io/projected/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-kube-api-access-5zm9g\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542884 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-ovs-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542915 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-nmstate-lock\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: E1124 13:56:43.542904 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair podName:892ed300-f0bb-49aa-80e3-b7feb492606c nodeName:}" failed. No retries permitted until 2025-11-24 13:56:44.042887176 +0000 UTC m=+619.330644469 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair") pod "nmstate-webhook-6b89b748d8-bjrv5" (UID: "892ed300-f0bb-49aa-80e3-b7feb492606c") : secret "openshift-nmstate-webhook" not found Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.542956 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.543097 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-dbus-socket\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.553025 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.572933 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g28h5\" (UniqueName: \"kubernetes.io/projected/892ed300-f0bb-49aa-80e3-b7feb492606c-kube-api-access-g28h5\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.584501 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5868b4846d-tcqsm"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.585364 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.592231 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zm9g\" (UniqueName: \"kubernetes.io/projected/2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48-kube-api-access-5zm9g\") pod \"nmstate-handler-fpsdv\" (UID: \"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48\") " pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.601192 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5868b4846d-tcqsm"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.623284 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.643739 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv6dr\" (UniqueName: \"kubernetes.io/projected/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-kube-api-access-xv6dr\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.643817 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.643857 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.644886 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.647625 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.660170 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv6dr\" (UniqueName: \"kubernetes.io/projected/ebda0f09-edf4-40b6-a13d-f04cf3ddd510-kube-api-access-xv6dr\") pod \"nmstate-console-plugin-5874bd7bc5-nwkjr\" (UID: \"ebda0f09-edf4-40b6-a13d-f04cf3ddd510\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.710512 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.744789 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.744852 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.744907 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjhz8\" (UniqueName: \"kubernetes.io/projected/87a47d3e-1d68-42ff-bba3-e891b0c8420f-kube-api-access-tjhz8\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.744979 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-oauth-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.745024 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-service-ca\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.745054 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-trusted-ca-bundle\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.745096 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-oauth-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846511 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-oauth-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846562 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-service-ca\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846601 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-trusted-ca-bundle\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846621 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-oauth-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846655 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846683 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.846700 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjhz8\" (UniqueName: \"kubernetes.io/projected/87a47d3e-1d68-42ff-bba3-e891b0c8420f-kube-api-access-tjhz8\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.847771 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-trusted-ca-bundle\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.847815 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-service-ca\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.847956 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.848165 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/87a47d3e-1d68-42ff-bba3-e891b0c8420f-oauth-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.850093 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-oauth-config\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.850308 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/87a47d3e-1d68-42ff-bba3-e891b0c8420f-console-serving-cert\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.864854 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjhz8\" (UniqueName: \"kubernetes.io/projected/87a47d3e-1d68-42ff-bba3-e891b0c8420f-kube-api-access-tjhz8\") pod \"console-5868b4846d-tcqsm\" (UID: \"87a47d3e-1d68-42ff-bba3-e891b0c8420f\") " pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.895619 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.964731 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4"] Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.965869 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fpsdv" event={"ID":"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48","Type":"ContainerStarted","Data":"4092a1f5cac90891778b366b7ea7f98946849402702326addf208f142b93ab06"} Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.967375 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:43 crc kubenswrapper[4970]: I1124 13:56:43.968749 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" event={"ID":"ebda0f09-edf4-40b6-a13d-f04cf3ddd510","Type":"ContainerStarted","Data":"667f4c6ab42c7260ad327b5b312a9d969288ecb3ab87aedcf679c4d37a732442"} Nov 24 13:56:43 crc kubenswrapper[4970]: W1124 13:56:43.971089 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62a1e00f_7e1a_4e97_a311_5628f78a7798.slice/crio-02097e99aa6256a767fdaede9ae204329ab65491b13af45a19201cb0e71b30fa WatchSource:0}: Error finding container 02097e99aa6256a767fdaede9ae204329ab65491b13af45a19201cb0e71b30fa: Status 404 returned error can't find the container with id 02097e99aa6256a767fdaede9ae204329ab65491b13af45a19201cb0e71b30fa Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.049534 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.055763 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/892ed300-f0bb-49aa-80e3-b7feb492606c-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-bjrv5\" (UID: \"892ed300-f0bb-49aa-80e3-b7feb492606c\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.176262 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.347047 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5868b4846d-tcqsm"] Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.350349 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5"] Nov 24 13:56:44 crc kubenswrapper[4970]: W1124 13:56:44.354736 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87a47d3e_1d68_42ff_bba3_e891b0c8420f.slice/crio-533a4afb4ccbf6b6a774a8bc23ad3b4db0dc0dc6a1e172601beffc2053b86c72 WatchSource:0}: Error finding container 533a4afb4ccbf6b6a774a8bc23ad3b4db0dc0dc6a1e172601beffc2053b86c72: Status 404 returned error can't find the container with id 533a4afb4ccbf6b6a774a8bc23ad3b4db0dc0dc6a1e172601beffc2053b86c72 Nov 24 13:56:44 crc kubenswrapper[4970]: W1124 13:56:44.357380 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod892ed300_f0bb_49aa_80e3_b7feb492606c.slice/crio-4e4632a4f408aef34fb7db2d01949b71fc0bf3a931fc21bb2da2ed53b718630c WatchSource:0}: Error finding container 4e4632a4f408aef34fb7db2d01949b71fc0bf3a931fc21bb2da2ed53b718630c: Status 404 returned error can't find the container with id 4e4632a4f408aef34fb7db2d01949b71fc0bf3a931fc21bb2da2ed53b718630c Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.976259 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5868b4846d-tcqsm" event={"ID":"87a47d3e-1d68-42ff-bba3-e891b0c8420f","Type":"ContainerStarted","Data":"e816491f38e32868847534378de1ef5c16afe249ac6ffe9a212e4cd4f62c0241"} Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.976722 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5868b4846d-tcqsm" event={"ID":"87a47d3e-1d68-42ff-bba3-e891b0c8420f","Type":"ContainerStarted","Data":"533a4afb4ccbf6b6a774a8bc23ad3b4db0dc0dc6a1e172601beffc2053b86c72"} Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.979051 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" event={"ID":"62a1e00f-7e1a-4e97-a311-5628f78a7798","Type":"ContainerStarted","Data":"02097e99aa6256a767fdaede9ae204329ab65491b13af45a19201cb0e71b30fa"} Nov 24 13:56:44 crc kubenswrapper[4970]: I1124 13:56:44.981540 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" event={"ID":"892ed300-f0bb-49aa-80e3-b7feb492606c","Type":"ContainerStarted","Data":"4e4632a4f408aef34fb7db2d01949b71fc0bf3a931fc21bb2da2ed53b718630c"} Nov 24 13:56:45 crc kubenswrapper[4970]: I1124 13:56:45.009064 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5868b4846d-tcqsm" podStartSLOduration=2.009035311 podStartE2EDuration="2.009035311s" podCreationTimestamp="2025-11-24 13:56:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:56:45.005299667 +0000 UTC m=+620.293056970" watchObservedRunningTime="2025-11-24 13:56:45.009035311 +0000 UTC m=+620.296792644" Nov 24 13:56:46 crc kubenswrapper[4970]: I1124 13:56:46.996411 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" event={"ID":"892ed300-f0bb-49aa-80e3-b7feb492606c","Type":"ContainerStarted","Data":"4f29fb5d592677721cb2bff797b352852e20c298134ddedf950813a162a5d80b"} Nov 24 13:56:46 crc kubenswrapper[4970]: I1124 13:56:46.997046 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:56:46 crc kubenswrapper[4970]: I1124 13:56:46.998281 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" event={"ID":"62a1e00f-7e1a-4e97-a311-5628f78a7798","Type":"ContainerStarted","Data":"dfc68abdd6578db04cf8646ec50b0adda9d9441a402a81c6300a3aa26508e3ea"} Nov 24 13:56:47 crc kubenswrapper[4970]: I1124 13:56:47.001026 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fpsdv" event={"ID":"2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48","Type":"ContainerStarted","Data":"94e7bf662762f00fc984d2c1865335ed5f411e5c05fa3904ea7ff8d4f4944f85"} Nov 24 13:56:47 crc kubenswrapper[4970]: I1124 13:56:47.001213 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:47 crc kubenswrapper[4970]: I1124 13:56:47.012193 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" podStartSLOduration=2.214521286 podStartE2EDuration="4.012179806s" podCreationTimestamp="2025-11-24 13:56:43 +0000 UTC" firstStartedPulling="2025-11-24 13:56:44.360286699 +0000 UTC m=+619.648043992" lastFinishedPulling="2025-11-24 13:56:46.157945179 +0000 UTC m=+621.445702512" observedRunningTime="2025-11-24 13:56:47.010864069 +0000 UTC m=+622.298621362" watchObservedRunningTime="2025-11-24 13:56:47.012179806 +0000 UTC m=+622.299937099" Nov 24 13:56:47 crc kubenswrapper[4970]: I1124 13:56:47.027161 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fpsdv" podStartSLOduration=1.514068688 podStartE2EDuration="4.027139441s" podCreationTimestamp="2025-11-24 13:56:43 +0000 UTC" firstStartedPulling="2025-11-24 13:56:43.641800051 +0000 UTC m=+618.929557344" lastFinishedPulling="2025-11-24 13:56:46.154870774 +0000 UTC m=+621.442628097" observedRunningTime="2025-11-24 13:56:47.024013985 +0000 UTC m=+622.311771298" watchObservedRunningTime="2025-11-24 13:56:47.027139441 +0000 UTC m=+622.314896744" Nov 24 13:56:48 crc kubenswrapper[4970]: I1124 13:56:48.008731 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" event={"ID":"ebda0f09-edf4-40b6-a13d-f04cf3ddd510","Type":"ContainerStarted","Data":"d8d643765ca1883b4f3fc2e2f73d028af3102b3e93b1b8fc284be9c40bc1141d"} Nov 24 13:56:49 crc kubenswrapper[4970]: I1124 13:56:49.025712 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" event={"ID":"62a1e00f-7e1a-4e97-a311-5628f78a7798","Type":"ContainerStarted","Data":"40ce7223902e5356377ac0b1eaefb0afa09e3bb842aacc3351a705e664394e3e"} Nov 24 13:56:49 crc kubenswrapper[4970]: I1124 13:56:49.050381 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-nwkjr" podStartSLOduration=2.684182036 podStartE2EDuration="6.050347453s" podCreationTimestamp="2025-11-24 13:56:43 +0000 UTC" firstStartedPulling="2025-11-24 13:56:43.90717742 +0000 UTC m=+619.194934713" lastFinishedPulling="2025-11-24 13:56:47.273342837 +0000 UTC m=+622.561100130" observedRunningTime="2025-11-24 13:56:48.025122589 +0000 UTC m=+623.312879882" watchObservedRunningTime="2025-11-24 13:56:49.050347453 +0000 UTC m=+624.338104786" Nov 24 13:56:53 crc kubenswrapper[4970]: I1124 13:56:53.647127 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fpsdv" Nov 24 13:56:53 crc kubenswrapper[4970]: I1124 13:56:53.665775 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8nmq4" podStartSLOduration=6.198913717 podStartE2EDuration="10.665748073s" podCreationTimestamp="2025-11-24 13:56:43 +0000 UTC" firstStartedPulling="2025-11-24 13:56:43.974488518 +0000 UTC m=+619.262245811" lastFinishedPulling="2025-11-24 13:56:48.441322874 +0000 UTC m=+623.729080167" observedRunningTime="2025-11-24 13:56:49.051191807 +0000 UTC m=+624.338949160" watchObservedRunningTime="2025-11-24 13:56:53.665748073 +0000 UTC m=+628.953505366" Nov 24 13:56:53 crc kubenswrapper[4970]: I1124 13:56:53.968644 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:53 crc kubenswrapper[4970]: I1124 13:56:53.968719 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:53 crc kubenswrapper[4970]: I1124 13:56:53.975449 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:54 crc kubenswrapper[4970]: I1124 13:56:54.070667 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5868b4846d-tcqsm" Nov 24 13:56:54 crc kubenswrapper[4970]: I1124 13:56:54.191277 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:57:04 crc kubenswrapper[4970]: I1124 13:57:04.182912 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-bjrv5" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.235054 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-fsgxr" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerName="console" containerID="cri-o://e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec" gracePeriod=15 Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.603014 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fsgxr_45d80403-6fe8-4874-a2d9-04a3838b65f1/console/0.log" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.603411 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.784984 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.785053 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.785113 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.785144 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.785981 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config" (OuterVolumeSpecName: "console-config") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.786240 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.786210 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca" (OuterVolumeSpecName: "service-ca") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.786836 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.786306 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msw6g\" (UniqueName: \"kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.786945 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle\") pod \"45d80403-6fe8-4874-a2d9-04a3838b65f1\" (UID: \"45d80403-6fe8-4874-a2d9-04a3838b65f1\") " Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.787491 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.787819 4970 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.787841 4970 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.787852 4970 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.787904 4970 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/45d80403-6fe8-4874-a2d9-04a3838b65f1-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.792719 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.793193 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.793674 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g" (OuterVolumeSpecName: "kube-api-access-msw6g") pod "45d80403-6fe8-4874-a2d9-04a3838b65f1" (UID: "45d80403-6fe8-4874-a2d9-04a3838b65f1"). InnerVolumeSpecName "kube-api-access-msw6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.889633 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msw6g\" (UniqueName: \"kubernetes.io/projected/45d80403-6fe8-4874-a2d9-04a3838b65f1-kube-api-access-msw6g\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.889703 4970 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.889735 4970 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/45d80403-6fe8-4874-a2d9-04a3838b65f1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.994022 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc"] Nov 24 13:57:19 crc kubenswrapper[4970]: E1124 13:57:19.994471 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerName="console" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.994488 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerName="console" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.994639 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerName="console" Nov 24 13:57:19 crc kubenswrapper[4970]: I1124 13:57:19.995408 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.000975 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.001746 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc"] Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.095419 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s25m2\" (UniqueName: \"kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.095946 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.096281 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.197416 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.197481 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.197517 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s25m2\" (UniqueName: \"kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.198568 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.198744 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.226729 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s25m2\" (UniqueName: \"kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.264757 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-fsgxr_45d80403-6fe8-4874-a2d9-04a3838b65f1/console/0.log" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.264854 4970 generic.go:334] "Generic (PLEG): container finished" podID="45d80403-6fe8-4874-a2d9-04a3838b65f1" containerID="e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec" exitCode=2 Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.264930 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-fsgxr" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.264916 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fsgxr" event={"ID":"45d80403-6fe8-4874-a2d9-04a3838b65f1","Type":"ContainerDied","Data":"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec"} Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.265178 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-fsgxr" event={"ID":"45d80403-6fe8-4874-a2d9-04a3838b65f1","Type":"ContainerDied","Data":"f59069565cf50d0421f0e89c30c44525efc64706625f8028bc457696b2f623fd"} Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.265246 4970 scope.go:117] "RemoveContainer" containerID="e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.293047 4970 scope.go:117] "RemoveContainer" containerID="e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec" Nov 24 13:57:20 crc kubenswrapper[4970]: E1124 13:57:20.293779 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec\": container with ID starting with e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec not found: ID does not exist" containerID="e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.293851 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec"} err="failed to get container status \"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec\": rpc error: code = NotFound desc = could not find container \"e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec\": container with ID starting with e2e00ab9a8fb00fb9f8b2e6466ee00158adb9a4abd7097c86888d61f36fee3ec not found: ID does not exist" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.308863 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.313490 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.316247 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-fsgxr"] Nov 24 13:57:20 crc kubenswrapper[4970]: I1124 13:57:20.571127 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc"] Nov 24 13:57:21 crc kubenswrapper[4970]: I1124 13:57:21.275336 4970 generic.go:334] "Generic (PLEG): container finished" podID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerID="f96a86c879137981140532a53c6cba3c3da89ea3f6040b409ef227486288bf4e" exitCode=0 Nov 24 13:57:21 crc kubenswrapper[4970]: I1124 13:57:21.275418 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" event={"ID":"a3958bff-b868-47ce-bd5d-7adb85d9dab0","Type":"ContainerDied","Data":"f96a86c879137981140532a53c6cba3c3da89ea3f6040b409ef227486288bf4e"} Nov 24 13:57:21 crc kubenswrapper[4970]: I1124 13:57:21.275748 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" event={"ID":"a3958bff-b868-47ce-bd5d-7adb85d9dab0","Type":"ContainerStarted","Data":"012c532214d5ec5f96eddd971570b8899f142f58e6334adc2aad3694b4ca0052"} Nov 24 13:57:21 crc kubenswrapper[4970]: I1124 13:57:21.480261 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d80403-6fe8-4874-a2d9-04a3838b65f1" path="/var/lib/kubelet/pods/45d80403-6fe8-4874-a2d9-04a3838b65f1/volumes" Nov 24 13:57:23 crc kubenswrapper[4970]: I1124 13:57:23.291263 4970 generic.go:334] "Generic (PLEG): container finished" podID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerID="10effd8f905af3f4e9913643a84e03a019c429bef9da0586d14d18289d269891" exitCode=0 Nov 24 13:57:23 crc kubenswrapper[4970]: I1124 13:57:23.291369 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" event={"ID":"a3958bff-b868-47ce-bd5d-7adb85d9dab0","Type":"ContainerDied","Data":"10effd8f905af3f4e9913643a84e03a019c429bef9da0586d14d18289d269891"} Nov 24 13:57:24 crc kubenswrapper[4970]: I1124 13:57:24.302885 4970 generic.go:334] "Generic (PLEG): container finished" podID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerID="6aab92589e1edda6a02d679c9827a99eebff61ec4b2ed9d5ff883b23aa67fa33" exitCode=0 Nov 24 13:57:24 crc kubenswrapper[4970]: I1124 13:57:24.303012 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" event={"ID":"a3958bff-b868-47ce-bd5d-7adb85d9dab0","Type":"ContainerDied","Data":"6aab92589e1edda6a02d679c9827a99eebff61ec4b2ed9d5ff883b23aa67fa33"} Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.635086 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.681593 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util\") pod \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.681641 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle\") pod \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.681890 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s25m2\" (UniqueName: \"kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2\") pod \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\" (UID: \"a3958bff-b868-47ce-bd5d-7adb85d9dab0\") " Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.682822 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle" (OuterVolumeSpecName: "bundle") pod "a3958bff-b868-47ce-bd5d-7adb85d9dab0" (UID: "a3958bff-b868-47ce-bd5d-7adb85d9dab0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.689665 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2" (OuterVolumeSpecName: "kube-api-access-s25m2") pod "a3958bff-b868-47ce-bd5d-7adb85d9dab0" (UID: "a3958bff-b868-47ce-bd5d-7adb85d9dab0"). InnerVolumeSpecName "kube-api-access-s25m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.695297 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util" (OuterVolumeSpecName: "util") pod "a3958bff-b868-47ce-bd5d-7adb85d9dab0" (UID: "a3958bff-b868-47ce-bd5d-7adb85d9dab0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.782864 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s25m2\" (UniqueName: \"kubernetes.io/projected/a3958bff-b868-47ce-bd5d-7adb85d9dab0-kube-api-access-s25m2\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.782907 4970 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:25 crc kubenswrapper[4970]: I1124 13:57:25.782929 4970 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a3958bff-b868-47ce-bd5d-7adb85d9dab0-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:57:26 crc kubenswrapper[4970]: I1124 13:57:26.319336 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" event={"ID":"a3958bff-b868-47ce-bd5d-7adb85d9dab0","Type":"ContainerDied","Data":"012c532214d5ec5f96eddd971570b8899f142f58e6334adc2aad3694b4ca0052"} Nov 24 13:57:26 crc kubenswrapper[4970]: I1124 13:57:26.319394 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="012c532214d5ec5f96eddd971570b8899f142f58e6334adc2aad3694b4ca0052" Nov 24 13:57:26 crc kubenswrapper[4970]: I1124 13:57:26.319483 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.573734 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7"] Nov 24 13:57:38 crc kubenswrapper[4970]: E1124 13:57:38.574522 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="extract" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.574538 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="extract" Nov 24 13:57:38 crc kubenswrapper[4970]: E1124 13:57:38.574551 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="util" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.574558 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="util" Nov 24 13:57:38 crc kubenswrapper[4970]: E1124 13:57:38.574592 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="pull" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.574601 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="pull" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.574721 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3958bff-b868-47ce-bd5d-7adb85d9dab0" containerName="extract" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.575170 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.580166 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.580166 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.580281 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.581365 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.582334 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-47zbw" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.592526 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7"] Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.681394 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-apiservice-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.681691 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-webhook-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.681737 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngzdt\" (UniqueName: \"kubernetes.io/projected/39678371-0ddb-40cb-8b7d-375d70868b75-kube-api-access-ngzdt\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.783341 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngzdt\" (UniqueName: \"kubernetes.io/projected/39678371-0ddb-40cb-8b7d-375d70868b75-kube-api-access-ngzdt\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.783403 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-apiservice-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.783478 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-webhook-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.793094 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-apiservice-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.806269 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/39678371-0ddb-40cb-8b7d-375d70868b75-webhook-cert\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.816489 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngzdt\" (UniqueName: \"kubernetes.io/projected/39678371-0ddb-40cb-8b7d-375d70868b75-kube-api-access-ngzdt\") pod \"metallb-operator-controller-manager-56999fcf9c-k9nw7\" (UID: \"39678371-0ddb-40cb-8b7d-375d70868b75\") " pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.825744 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn"] Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.826759 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.828303 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-kwg44" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.830027 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.830361 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.843011 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn"] Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.884727 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-apiservice-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.884801 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6d7f\" (UniqueName: \"kubernetes.io/projected/db2d684a-3178-4375-b857-fc4ab433a328-kube-api-access-w6d7f\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.884831 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-webhook-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.893325 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.985909 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-apiservice-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.985993 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6d7f\" (UniqueName: \"kubernetes.io/projected/db2d684a-3178-4375-b857-fc4ab433a328-kube-api-access-w6d7f\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.986031 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-webhook-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:38 crc kubenswrapper[4970]: I1124 13:57:38.992199 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-apiservice-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.000487 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/db2d684a-3178-4375-b857-fc4ab433a328-webhook-cert\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.006188 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6d7f\" (UniqueName: \"kubernetes.io/projected/db2d684a-3178-4375-b857-fc4ab433a328-kube-api-access-w6d7f\") pod \"metallb-operator-webhook-server-56df5758cc-gjrmn\" (UID: \"db2d684a-3178-4375-b857-fc4ab433a328\") " pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.160006 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7"] Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.175329 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:39 crc kubenswrapper[4970]: W1124 13:57:39.179955 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39678371_0ddb_40cb_8b7d_375d70868b75.slice/crio-ce824b5e7de2e67005bfe8ada7d14234831069591af3d65ecfb259e549c2f22f WatchSource:0}: Error finding container ce824b5e7de2e67005bfe8ada7d14234831069591af3d65ecfb259e549c2f22f: Status 404 returned error can't find the container with id ce824b5e7de2e67005bfe8ada7d14234831069591af3d65ecfb259e549c2f22f Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.400430 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" event={"ID":"39678371-0ddb-40cb-8b7d-375d70868b75","Type":"ContainerStarted","Data":"ce824b5e7de2e67005bfe8ada7d14234831069591af3d65ecfb259e549c2f22f"} Nov 24 13:57:39 crc kubenswrapper[4970]: I1124 13:57:39.414934 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn"] Nov 24 13:57:39 crc kubenswrapper[4970]: W1124 13:57:39.419919 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb2d684a_3178_4375_b857_fc4ab433a328.slice/crio-0c5f49ab5da6867554d140e9f1f4fa2f1b789591a470bb43154f7f5c90bd7c13 WatchSource:0}: Error finding container 0c5f49ab5da6867554d140e9f1f4fa2f1b789591a470bb43154f7f5c90bd7c13: Status 404 returned error can't find the container with id 0c5f49ab5da6867554d140e9f1f4fa2f1b789591a470bb43154f7f5c90bd7c13 Nov 24 13:57:40 crc kubenswrapper[4970]: I1124 13:57:40.425906 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" event={"ID":"db2d684a-3178-4375-b857-fc4ab433a328","Type":"ContainerStarted","Data":"0c5f49ab5da6867554d140e9f1f4fa2f1b789591a470bb43154f7f5c90bd7c13"} Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.466323 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" event={"ID":"db2d684a-3178-4375-b857-fc4ab433a328","Type":"ContainerStarted","Data":"2f5d99cc1036233ddaf0f337833b527da182fcde13d6354242b74a720bf34402"} Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.466892 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.467949 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" event={"ID":"39678371-0ddb-40cb-8b7d-375d70868b75","Type":"ContainerStarted","Data":"7cb9b9e99370308c83548811a00ae0c0b82d5938b17cb36d3223739ea27f0b6a"} Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.468039 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.485062 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" podStartSLOduration=2.480685429 podStartE2EDuration="6.485040699s" podCreationTimestamp="2025-11-24 13:57:38 +0000 UTC" firstStartedPulling="2025-11-24 13:57:39.423265945 +0000 UTC m=+674.711023238" lastFinishedPulling="2025-11-24 13:57:43.427621215 +0000 UTC m=+678.715378508" observedRunningTime="2025-11-24 13:57:44.481897473 +0000 UTC m=+679.769654786" watchObservedRunningTime="2025-11-24 13:57:44.485040699 +0000 UTC m=+679.772797992" Nov 24 13:57:44 crc kubenswrapper[4970]: I1124 13:57:44.509643 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" podStartSLOduration=2.283599302 podStartE2EDuration="6.509619651s" podCreationTimestamp="2025-11-24 13:57:38 +0000 UTC" firstStartedPulling="2025-11-24 13:57:39.192834816 +0000 UTC m=+674.480592109" lastFinishedPulling="2025-11-24 13:57:43.418855155 +0000 UTC m=+678.706612458" observedRunningTime="2025-11-24 13:57:44.507041021 +0000 UTC m=+679.794798324" watchObservedRunningTime="2025-11-24 13:57:44.509619651 +0000 UTC m=+679.797376944" Nov 24 13:57:59 crc kubenswrapper[4970]: I1124 13:57:59.186968 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-56df5758cc-gjrmn" Nov 24 13:58:11 crc kubenswrapper[4970]: I1124 13:58:11.204831 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:58:11 crc kubenswrapper[4970]: I1124 13:58:11.205529 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:58:18 crc kubenswrapper[4970]: I1124 13:58:18.895941 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-56999fcf9c-k9nw7" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.825460 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-vgst9"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.853745 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-c767v"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.854907 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.855039 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.860333 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.860497 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.860686 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rn4p9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.861933 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-sockets\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.861978 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics-certs\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862004 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-reloader\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862028 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt75z\" (UniqueName: \"kubernetes.io/projected/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-kube-api-access-qt75z\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862048 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-cert\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862319 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-startup\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862397 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862422 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-conf\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.862441 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tw2b\" (UniqueName: \"kubernetes.io/projected/9bd50032-57ea-42cc-9c4c-0a4213b02636-kube-api-access-7tw2b\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.869082 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.873771 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-c767v"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.912396 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qn5c2"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.913220 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qn5c2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.920279 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.920846 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qflgl" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.920309 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.920493 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.950740 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-s8fl4"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.951651 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.956555 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-s8fl4"] Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.958990 4970 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.963649 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.963862 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.963966 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-cert\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964059 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2bc53266-821b-4f2b-af20-c290fff7e4a9-metallb-excludel2\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964218 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964326 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-conf\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964431 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tw2b\" (UniqueName: \"kubernetes.io/projected/9bd50032-57ea-42cc-9c4c-0a4213b02636-kube-api-access-7tw2b\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964539 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lqqf\" (UniqueName: \"kubernetes.io/projected/48658fb0-3b0d-4505-9e99-68d21fa216e8-kube-api-access-6lqqf\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964681 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-sockets\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.964922 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-metrics-certs\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965027 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics-certs\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965117 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-conf\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965139 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965129 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-reloader\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965235 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-cert\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965263 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt75z\" (UniqueName: \"kubernetes.io/projected/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-kube-api-access-qt75z\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965301 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-startup\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965518 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc6bn\" (UniqueName: \"kubernetes.io/projected/2bc53266-821b-4f2b-af20-c290fff7e4a9-kube-api-access-lc6bn\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.965961 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-sockets\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.966287 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/9bd50032-57ea-42cc-9c4c-0a4213b02636-reloader\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.966561 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/9bd50032-57ea-42cc-9c4c-0a4213b02636-frr-startup\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.976527 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-cert\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.978117 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9bd50032-57ea-42cc-9c4c-0a4213b02636-metrics-certs\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.989059 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tw2b\" (UniqueName: \"kubernetes.io/projected/9bd50032-57ea-42cc-9c4c-0a4213b02636-kube-api-access-7tw2b\") pod \"frr-k8s-vgst9\" (UID: \"9bd50032-57ea-42cc-9c4c-0a4213b02636\") " pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:19 crc kubenswrapper[4970]: I1124 13:58:19.991880 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt75z\" (UniqueName: \"kubernetes.io/projected/fdf01015-8aee-4c8d-bdd2-0a72e6387d32-kube-api-access-qt75z\") pod \"frr-k8s-webhook-server-6998585d5-c767v\" (UID: \"fdf01015-8aee-4c8d-bdd2-0a72e6387d32\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066740 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc6bn\" (UniqueName: \"kubernetes.io/projected/2bc53266-821b-4f2b-af20-c290fff7e4a9-kube-api-access-lc6bn\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066798 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066825 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066847 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-cert\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066864 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2bc53266-821b-4f2b-af20-c290fff7e4a9-metallb-excludel2\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066894 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lqqf\" (UniqueName: \"kubernetes.io/projected/48658fb0-3b0d-4505-9e99-68d21fa216e8-kube-api-access-6lqqf\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.066933 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-metrics-certs\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.067441 4970 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.067515 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist podName:2bc53266-821b-4f2b-af20-c290fff7e4a9 nodeName:}" failed. No retries permitted until 2025-11-24 13:58:20.567498285 +0000 UTC m=+715.855255578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist") pod "speaker-qn5c2" (UID: "2bc53266-821b-4f2b-af20-c290fff7e4a9") : secret "metallb-memberlist" not found Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.067896 4970 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.067927 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs podName:48658fb0-3b0d-4505-9e99-68d21fa216e8 nodeName:}" failed. No retries permitted until 2025-11-24 13:58:20.567920416 +0000 UTC m=+715.855677709 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs") pod "controller-6c7b4b5f48-s8fl4" (UID: "48658fb0-3b0d-4505-9e99-68d21fa216e8") : secret "controller-certs-secret" not found Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.068450 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/2bc53266-821b-4f2b-af20-c290fff7e4a9-metallb-excludel2\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.069921 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-metrics-certs\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.071027 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-cert\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.082081 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc6bn\" (UniqueName: \"kubernetes.io/projected/2bc53266-821b-4f2b-af20-c290fff7e4a9-kube-api-access-lc6bn\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.082779 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lqqf\" (UniqueName: \"kubernetes.io/projected/48658fb0-3b0d-4505-9e99-68d21fa216e8-kube-api-access-6lqqf\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.186723 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.204010 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.480294 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-c767v"] Nov 24 13:58:20 crc kubenswrapper[4970]: W1124 13:58:20.485381 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdf01015_8aee_4c8d_bdd2_0a72e6387d32.slice/crio-93d11b4504ceb0db139664e97a4cc01dd6c38fe1338d17cfa1d44253c06c1f3f WatchSource:0}: Error finding container 93d11b4504ceb0db139664e97a4cc01dd6c38fe1338d17cfa1d44253c06c1f3f: Status 404 returned error can't find the container with id 93d11b4504ceb0db139664e97a4cc01dd6c38fe1338d17cfa1d44253c06c1f3f Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.576151 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.576195 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.576313 4970 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 13:58:20 crc kubenswrapper[4970]: E1124 13:58:20.576357 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist podName:2bc53266-821b-4f2b-af20-c290fff7e4a9 nodeName:}" failed. No retries permitted until 2025-11-24 13:58:21.576344033 +0000 UTC m=+716.864101326 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist") pod "speaker-qn5c2" (UID: "2bc53266-821b-4f2b-af20-c290fff7e4a9") : secret "metallb-memberlist" not found Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.582432 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48658fb0-3b0d-4505-9e99-68d21fa216e8-metrics-certs\") pod \"controller-6c7b4b5f48-s8fl4\" (UID: \"48658fb0-3b0d-4505-9e99-68d21fa216e8\") " pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.682299 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" event={"ID":"fdf01015-8aee-4c8d-bdd2-0a72e6387d32","Type":"ContainerStarted","Data":"93d11b4504ceb0db139664e97a4cc01dd6c38fe1338d17cfa1d44253c06c1f3f"} Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.683336 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"cb16b82ca219f297267726d0affa74d4892a566c682c4a85a3ba55c619f99105"} Nov 24 13:58:20 crc kubenswrapper[4970]: I1124 13:58:20.875154 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.053088 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-s8fl4"] Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.597425 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.602249 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/2bc53266-821b-4f2b-af20-c290fff7e4a9-memberlist\") pod \"speaker-qn5c2\" (UID: \"2bc53266-821b-4f2b-af20-c290fff7e4a9\") " pod="metallb-system/speaker-qn5c2" Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.691895 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-s8fl4" event={"ID":"48658fb0-3b0d-4505-9e99-68d21fa216e8","Type":"ContainerStarted","Data":"31b99bc131d2d6ea9712bfaf49a89877f7a914950c543bb921ffc78950f95ca3"} Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.691939 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-s8fl4" event={"ID":"48658fb0-3b0d-4505-9e99-68d21fa216e8","Type":"ContainerStarted","Data":"8498dd2cd8d703df2dacfe93a4ce6d58b81ef15efe86abd8129b9c1583af4ece"} Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.691948 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-s8fl4" event={"ID":"48658fb0-3b0d-4505-9e99-68d21fa216e8","Type":"ContainerStarted","Data":"2511608f12a899b62153002974647fcc9d3ab1ac794d2dfe3cc975bc4c92862b"} Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.692040 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.709022 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-s8fl4" podStartSLOduration=2.709005041 podStartE2EDuration="2.709005041s" podCreationTimestamp="2025-11-24 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:58:21.707492752 +0000 UTC m=+716.995250045" watchObservedRunningTime="2025-11-24 13:58:21.709005041 +0000 UTC m=+716.996762354" Nov 24 13:58:21 crc kubenswrapper[4970]: I1124 13:58:21.735798 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qn5c2" Nov 24 13:58:21 crc kubenswrapper[4970]: W1124 13:58:21.753049 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bc53266_821b_4f2b_af20_c290fff7e4a9.slice/crio-42ee81459cc6c18f7798f419bc6e52a6276bf9db59435ce122223c9ff947c7d3 WatchSource:0}: Error finding container 42ee81459cc6c18f7798f419bc6e52a6276bf9db59435ce122223c9ff947c7d3: Status 404 returned error can't find the container with id 42ee81459cc6c18f7798f419bc6e52a6276bf9db59435ce122223c9ff947c7d3 Nov 24 13:58:22 crc kubenswrapper[4970]: I1124 13:58:22.705419 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qn5c2" event={"ID":"2bc53266-821b-4f2b-af20-c290fff7e4a9","Type":"ContainerStarted","Data":"a3bb58f8d629162abbaa9aa9d4dfba4ac2a42fbd614828e6aacb1bee5ea5fd02"} Nov 24 13:58:22 crc kubenswrapper[4970]: I1124 13:58:22.705765 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qn5c2" event={"ID":"2bc53266-821b-4f2b-af20-c290fff7e4a9","Type":"ContainerStarted","Data":"ffb8a3d5745e877af5c2e477ea9c31d2a65caebf9370f18f63b08f951b0ca116"} Nov 24 13:58:22 crc kubenswrapper[4970]: I1124 13:58:22.705784 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qn5c2" event={"ID":"2bc53266-821b-4f2b-af20-c290fff7e4a9","Type":"ContainerStarted","Data":"42ee81459cc6c18f7798f419bc6e52a6276bf9db59435ce122223c9ff947c7d3"} Nov 24 13:58:22 crc kubenswrapper[4970]: I1124 13:58:22.705959 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qn5c2" Nov 24 13:58:22 crc kubenswrapper[4970]: I1124 13:58:22.722878 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qn5c2" podStartSLOduration=3.722858767 podStartE2EDuration="3.722858767s" podCreationTimestamp="2025-11-24 13:58:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:58:22.722257981 +0000 UTC m=+718.010015274" watchObservedRunningTime="2025-11-24 13:58:22.722858767 +0000 UTC m=+718.010616060" Nov 24 13:58:27 crc kubenswrapper[4970]: I1124 13:58:27.733652 4970 generic.go:334] "Generic (PLEG): container finished" podID="9bd50032-57ea-42cc-9c4c-0a4213b02636" containerID="fe5c6ddbb174b7142192de8a5ea8ba8a32229e5106d638121b217cd026f14a4e" exitCode=0 Nov 24 13:58:27 crc kubenswrapper[4970]: I1124 13:58:27.733713 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerDied","Data":"fe5c6ddbb174b7142192de8a5ea8ba8a32229e5106d638121b217cd026f14a4e"} Nov 24 13:58:27 crc kubenswrapper[4970]: I1124 13:58:27.735490 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" event={"ID":"fdf01015-8aee-4c8d-bdd2-0a72e6387d32","Type":"ContainerStarted","Data":"095c71399f5fd41396232c497782a271d9b0e437eb31f3cb5a6cdc41f535814a"} Nov 24 13:58:27 crc kubenswrapper[4970]: I1124 13:58:27.735736 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:27 crc kubenswrapper[4970]: I1124 13:58:27.779325 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" podStartSLOduration=1.8125774730000002 podStartE2EDuration="8.77930259s" podCreationTimestamp="2025-11-24 13:58:19 +0000 UTC" firstStartedPulling="2025-11-24 13:58:20.487773827 +0000 UTC m=+715.775531120" lastFinishedPulling="2025-11-24 13:58:27.454498924 +0000 UTC m=+722.742256237" observedRunningTime="2025-11-24 13:58:27.77699069 +0000 UTC m=+723.064747983" watchObservedRunningTime="2025-11-24 13:58:27.77930259 +0000 UTC m=+723.067059883" Nov 24 13:58:28 crc kubenswrapper[4970]: I1124 13:58:28.746671 4970 generic.go:334] "Generic (PLEG): container finished" podID="9bd50032-57ea-42cc-9c4c-0a4213b02636" containerID="2020d5652a1e550c4d009758a2123173094db2680d18634f4fcc500e38b9a635" exitCode=0 Nov 24 13:58:28 crc kubenswrapper[4970]: I1124 13:58:28.746827 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerDied","Data":"2020d5652a1e550c4d009758a2123173094db2680d18634f4fcc500e38b9a635"} Nov 24 13:58:29 crc kubenswrapper[4970]: I1124 13:58:29.755853 4970 generic.go:334] "Generic (PLEG): container finished" podID="9bd50032-57ea-42cc-9c4c-0a4213b02636" containerID="5ac705696fc706aac76ede13ae91a126b89474fabae6002663685498f882e808" exitCode=0 Nov 24 13:58:29 crc kubenswrapper[4970]: I1124 13:58:29.756106 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerDied","Data":"5ac705696fc706aac76ede13ae91a126b89474fabae6002663685498f882e808"} Nov 24 13:58:30 crc kubenswrapper[4970]: I1124 13:58:30.768566 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"d3ab7b88cf78f51f775d69a0ac0abdb222efae0dc106a35d11a91ceb352b5b5c"} Nov 24 13:58:30 crc kubenswrapper[4970]: I1124 13:58:30.768920 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"03b86b0dec4f564eec295b40627419e0ba13f27cd012d77e73bd0824aeb0ad4e"} Nov 24 13:58:30 crc kubenswrapper[4970]: I1124 13:58:30.768937 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"ccf1b5390ab2d21fab5ae51af8026f912dfb3af571598ba8ba1fb9f5151a9229"} Nov 24 13:58:30 crc kubenswrapper[4970]: I1124 13:58:30.768949 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"53c41bbe81bee0c557e825b1dd3cba4fed0b1ce2234692fad69fe08784452c0c"} Nov 24 13:58:30 crc kubenswrapper[4970]: I1124 13:58:30.768960 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"23a7c938dc5136759eb5071ccf9be95f62df37d3418c8183f8f9b8668e4f9f8c"} Nov 24 13:58:31 crc kubenswrapper[4970]: I1124 13:58:31.778234 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vgst9" event={"ID":"9bd50032-57ea-42cc-9c4c-0a4213b02636","Type":"ContainerStarted","Data":"edb52717def57d0b66f51990503c4803991ed992d0d96e3bfd72766eab887fec"} Nov 24 13:58:31 crc kubenswrapper[4970]: I1124 13:58:31.779191 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:35 crc kubenswrapper[4970]: I1124 13:58:35.187640 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:35 crc kubenswrapper[4970]: I1124 13:58:35.251043 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:35 crc kubenswrapper[4970]: I1124 13:58:35.275027 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-vgst9" podStartSLOduration=9.141036648 podStartE2EDuration="16.27501249s" podCreationTimestamp="2025-11-24 13:58:19 +0000 UTC" firstStartedPulling="2025-11-24 13:58:20.35233824 +0000 UTC m=+715.640095553" lastFinishedPulling="2025-11-24 13:58:27.486314092 +0000 UTC m=+722.774071395" observedRunningTime="2025-11-24 13:58:31.802301769 +0000 UTC m=+727.090059072" watchObservedRunningTime="2025-11-24 13:58:35.27501249 +0000 UTC m=+730.562769783" Nov 24 13:58:40 crc kubenswrapper[4970]: I1124 13:58:40.190862 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-vgst9" Nov 24 13:58:40 crc kubenswrapper[4970]: I1124 13:58:40.217612 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-c767v" Nov 24 13:58:40 crc kubenswrapper[4970]: I1124 13:58:40.879493 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-s8fl4" Nov 24 13:58:41 crc kubenswrapper[4970]: I1124 13:58:41.204508 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:58:41 crc kubenswrapper[4970]: I1124 13:58:41.204620 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:58:41 crc kubenswrapper[4970]: I1124 13:58:41.740512 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qn5c2" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.191717 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.193414 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.195623 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8k9nj" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.197159 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.198335 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.205521 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.278324 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7r4d\" (UniqueName: \"kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d\") pod \"openstack-operator-index-zd5kp\" (UID: \"8f72328f-ea0c-4701-bd24-1e4cdf149962\") " pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.379784 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7r4d\" (UniqueName: \"kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d\") pod \"openstack-operator-index-zd5kp\" (UID: \"8f72328f-ea0c-4701-bd24-1e4cdf149962\") " pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.402394 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7r4d\" (UniqueName: \"kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d\") pod \"openstack-operator-index-zd5kp\" (UID: \"8f72328f-ea0c-4701-bd24-1e4cdf149962\") " pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.550838 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.786154 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:48 crc kubenswrapper[4970]: W1124 13:58:48.798978 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f72328f_ea0c_4701_bd24_1e4cdf149962.slice/crio-037c0fe1c593d85c1b51f85044413284ac34a3fd70d39ea26f1d2d216aa4459b WatchSource:0}: Error finding container 037c0fe1c593d85c1b51f85044413284ac34a3fd70d39ea26f1d2d216aa4459b: Status 404 returned error can't find the container with id 037c0fe1c593d85c1b51f85044413284ac34a3fd70d39ea26f1d2d216aa4459b Nov 24 13:58:48 crc kubenswrapper[4970]: I1124 13:58:48.901091 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zd5kp" event={"ID":"8f72328f-ea0c-4701-bd24-1e4cdf149962","Type":"ContainerStarted","Data":"037c0fe1c593d85c1b51f85044413284ac34a3fd70d39ea26f1d2d216aa4459b"} Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.050147 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.050385 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerName="controller-manager" containerID="cri-o://e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c" gracePeriod=30 Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.149153 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.149601 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerName="route-controller-manager" containerID="cri-o://ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b" gracePeriod=30 Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.506851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.632749 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.700434 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ntcq\" (UniqueName: \"kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq\") pod \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.700772 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config\") pod \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.700806 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert\") pod \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.700832 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles\") pod \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.700853 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca\") pod \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\" (UID: \"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.701367 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" (UID: "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.701571 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config" (OuterVolumeSpecName: "config") pod "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" (UID: "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.702679 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca" (OuterVolumeSpecName: "client-ca") pod "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" (UID: "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.705846 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" (UID: "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.706401 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq" (OuterVolumeSpecName: "kube-api-access-9ntcq") pod "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" (UID: "b7f2f7bc-97c0-4e99-8a46-32815fa43b5b"). InnerVolumeSpecName "kube-api-access-9ntcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802046 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2knwk\" (UniqueName: \"kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk\") pod \"861c2ae4-caff-41f9-a437-f7702a7a5c25\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802178 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config\") pod \"861c2ae4-caff-41f9-a437-f7702a7a5c25\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802214 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca\") pod \"861c2ae4-caff-41f9-a437-f7702a7a5c25\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802240 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert\") pod \"861c2ae4-caff-41f9-a437-f7702a7a5c25\" (UID: \"861c2ae4-caff-41f9-a437-f7702a7a5c25\") " Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802500 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ntcq\" (UniqueName: \"kubernetes.io/projected/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-kube-api-access-9ntcq\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802512 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802523 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802532 4970 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.802540 4970 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.803544 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config" (OuterVolumeSpecName: "config") pod "861c2ae4-caff-41f9-a437-f7702a7a5c25" (UID: "861c2ae4-caff-41f9-a437-f7702a7a5c25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.803855 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca" (OuterVolumeSpecName: "client-ca") pod "861c2ae4-caff-41f9-a437-f7702a7a5c25" (UID: "861c2ae4-caff-41f9-a437-f7702a7a5c25"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.806261 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "861c2ae4-caff-41f9-a437-f7702a7a5c25" (UID: "861c2ae4-caff-41f9-a437-f7702a7a5c25"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.809055 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk" (OuterVolumeSpecName: "kube-api-access-2knwk") pod "861c2ae4-caff-41f9-a437-f7702a7a5c25" (UID: "861c2ae4-caff-41f9-a437-f7702a7a5c25"). InnerVolumeSpecName "kube-api-access-2knwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.903402 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-config\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.903446 4970 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/861c2ae4-caff-41f9-a437-f7702a7a5c25-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.903456 4970 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/861c2ae4-caff-41f9-a437-f7702a7a5c25-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.903467 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2knwk\" (UniqueName: \"kubernetes.io/projected/861c2ae4-caff-41f9-a437-f7702a7a5c25-kube-api-access-2knwk\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.910679 4970 generic.go:334] "Generic (PLEG): container finished" podID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerID="e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c" exitCode=0 Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.910753 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" event={"ID":"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b","Type":"ContainerDied","Data":"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c"} Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.910780 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" event={"ID":"b7f2f7bc-97c0-4e99-8a46-32815fa43b5b","Type":"ContainerDied","Data":"201ced5d4f98e6653ae6fd0a1b92a36d6b365612a2c93a33cdcef205514dcc5d"} Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.910802 4970 scope.go:117] "RemoveContainer" containerID="e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.910851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rnm9t" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.912727 4970 generic.go:334] "Generic (PLEG): container finished" podID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerID="ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b" exitCode=0 Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.912758 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" event={"ID":"861c2ae4-caff-41f9-a437-f7702a7a5c25","Type":"ContainerDied","Data":"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b"} Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.912773 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" event={"ID":"861c2ae4-caff-41f9-a437-f7702a7a5c25","Type":"ContainerDied","Data":"5d094152c29dc4d8a45da75324864492547dd0f30ccdc79e17dc791714f9a0d8"} Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.912808 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.957502 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.959743 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c925s"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.968146 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.971273 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rnm9t"] Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.994118 4970 scope.go:117] "RemoveContainer" containerID="e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c" Nov 24 13:58:49 crc kubenswrapper[4970]: E1124 13:58:49.994654 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c\": container with ID starting with e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c not found: ID does not exist" containerID="e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.994734 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c"} err="failed to get container status \"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c\": rpc error: code = NotFound desc = could not find container \"e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c\": container with ID starting with e7f753b108d27a1f06d1923e8eab58683db02801a0ce59adc9a3c9752639af1c not found: ID does not exist" Nov 24 13:58:49 crc kubenswrapper[4970]: I1124 13:58:49.994781 4970 scope.go:117] "RemoveContainer" containerID="ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.024429 4970 scope.go:117] "RemoveContainer" containerID="ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b" Nov 24 13:58:50 crc kubenswrapper[4970]: E1124 13:58:50.025952 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b\": container with ID starting with ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b not found: ID does not exist" containerID="ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.026009 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b"} err="failed to get container status \"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b\": rpc error: code = NotFound desc = could not find container \"ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b\": container with ID starting with ec3c7f8d181a4905e0541f0d78c10e81454311a763e3a7cade96cfd2c76c642b not found: ID does not exist" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.842067 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq"] Nov 24 13:58:50 crc kubenswrapper[4970]: E1124 13:58:50.842626 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerName="controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.842642 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerName="controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: E1124 13:58:50.842664 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerName="route-controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.842672 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerName="route-controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.842809 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" containerName="route-controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.842825 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" containerName="controller-manager" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.843270 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.846748 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.846898 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.846747 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.846851 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.847337 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.853977 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq"] Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.856257 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.919386 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-config\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.919673 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-client-ca\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.919822 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e061390e-d5ae-432f-97bc-6064432611b6-serving-cert\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.920000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtd99\" (UniqueName: \"kubernetes.io/projected/e061390e-d5ae-432f-97bc-6064432611b6-kube-api-access-jtd99\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.920502 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zd5kp" event={"ID":"8f72328f-ea0c-4701-bd24-1e4cdf149962","Type":"ContainerStarted","Data":"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8"} Nov 24 13:58:50 crc kubenswrapper[4970]: I1124 13:58:50.939666 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zd5kp" podStartSLOduration=1.74557691 podStartE2EDuration="2.939647987s" podCreationTimestamp="2025-11-24 13:58:48 +0000 UTC" firstStartedPulling="2025-11-24 13:58:48.80184716 +0000 UTC m=+744.089604463" lastFinishedPulling="2025-11-24 13:58:49.995918237 +0000 UTC m=+745.283675540" observedRunningTime="2025-11-24 13:58:50.9382119 +0000 UTC m=+746.225969203" watchObservedRunningTime="2025-11-24 13:58:50.939647987 +0000 UTC m=+746.227405280" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.021040 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-config\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.021376 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-client-ca\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.021512 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e061390e-d5ae-432f-97bc-6064432611b6-serving-cert\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.021686 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtd99\" (UniqueName: \"kubernetes.io/projected/e061390e-d5ae-432f-97bc-6064432611b6-kube-api-access-jtd99\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.022774 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-client-ca\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.024801 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e061390e-d5ae-432f-97bc-6064432611b6-config\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.032103 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e061390e-d5ae-432f-97bc-6064432611b6-serving-cert\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.040758 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtd99\" (UniqueName: \"kubernetes.io/projected/e061390e-d5ae-432f-97bc-6064432611b6-kube-api-access-jtd99\") pod \"route-controller-manager-7fc58f478-b9gkq\" (UID: \"e061390e-d5ae-432f-97bc-6064432611b6\") " pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.093405 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2"] Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.096470 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.100440 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.100753 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.100796 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.100901 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.100963 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.101686 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.105768 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2"] Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.111673 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.160436 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.226613 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-proxy-ca-bundles\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.226981 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-config\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.227032 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a86a4c5-045f-48fc-875a-32037bccbcd5-serving-cert\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.227084 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2jpq\" (UniqueName: \"kubernetes.io/projected/3a86a4c5-045f-48fc-875a-32037bccbcd5-kube-api-access-t2jpq\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.227126 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-client-ca\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.328057 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a86a4c5-045f-48fc-875a-32037bccbcd5-serving-cert\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.328128 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2jpq\" (UniqueName: \"kubernetes.io/projected/3a86a4c5-045f-48fc-875a-32037bccbcd5-kube-api-access-t2jpq\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.328165 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-client-ca\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.328237 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-proxy-ca-bundles\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.328273 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-config\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.330504 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-client-ca\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.331264 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-config\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.333546 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a86a4c5-045f-48fc-875a-32037bccbcd5-proxy-ca-bundles\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.334274 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a86a4c5-045f-48fc-875a-32037bccbcd5-serving-cert\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.351462 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2jpq\" (UniqueName: \"kubernetes.io/projected/3a86a4c5-045f-48fc-875a-32037bccbcd5-kube-api-access-t2jpq\") pod \"controller-manager-6d6d7486bd-xp9d2\" (UID: \"3a86a4c5-045f-48fc-875a-32037bccbcd5\") " pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.387254 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq"] Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.420858 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.478842 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861c2ae4-caff-41f9-a437-f7702a7a5c25" path="/var/lib/kubelet/pods/861c2ae4-caff-41f9-a437-f7702a7a5c25/volumes" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.479729 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f2f7bc-97c0-4e99-8a46-32815fa43b5b" path="/var/lib/kubelet/pods/b7f2f7bc-97c0-4e99-8a46-32815fa43b5b/volumes" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.844257 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2"] Nov 24 13:58:51 crc kubenswrapper[4970]: W1124 13:58:51.853003 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a86a4c5_045f_48fc_875a_32037bccbcd5.slice/crio-cbc4b81d587a0304b889757fbc92d7bbc3fd804de78fe40837ded38f792f6d11 WatchSource:0}: Error finding container cbc4b81d587a0304b889757fbc92d7bbc3fd804de78fe40837ded38f792f6d11: Status 404 returned error can't find the container with id cbc4b81d587a0304b889757fbc92d7bbc3fd804de78fe40837ded38f792f6d11 Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.934050 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" event={"ID":"e061390e-d5ae-432f-97bc-6064432611b6","Type":"ContainerStarted","Data":"1a8f7c0d2fb45417af09f3a37a86c3ef009f8b6af17a394e4f41395a0023bcf8"} Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.934089 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" event={"ID":"e061390e-d5ae-432f-97bc-6064432611b6","Type":"ContainerStarted","Data":"7f554cce38c21c6f8c6988da2072ab8a63944dc29cbcaa705e1ce97d27a8b460"} Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.934278 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.935311 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" event={"ID":"3a86a4c5-045f-48fc-875a-32037bccbcd5","Type":"ContainerStarted","Data":"cbc4b81d587a0304b889757fbc92d7bbc3fd804de78fe40837ded38f792f6d11"} Nov 24 13:58:51 crc kubenswrapper[4970]: I1124 13:58:51.955514 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" podStartSLOduration=1.955499994 podStartE2EDuration="1.955499994s" podCreationTimestamp="2025-11-24 13:58:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:58:51.953277707 +0000 UTC m=+747.241034990" watchObservedRunningTime="2025-11-24 13:58:51.955499994 +0000 UTC m=+747.243257287" Nov 24 13:58:52 crc kubenswrapper[4970]: I1124 13:58:51.999700 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fc58f478-b9gkq" Nov 24 13:58:52 crc kubenswrapper[4970]: I1124 13:58:52.944566 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" event={"ID":"3a86a4c5-045f-48fc-875a-32037bccbcd5","Type":"ContainerStarted","Data":"58ac0c3df204e2b552676ddee548ea4f55c332698cb4040d7fde169f2e73a070"} Nov 24 13:58:52 crc kubenswrapper[4970]: I1124 13:58:52.944867 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:52 crc kubenswrapper[4970]: I1124 13:58:52.950058 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" Nov 24 13:58:52 crc kubenswrapper[4970]: I1124 13:58:52.966119 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d6d7486bd-xp9d2" podStartSLOduration=3.966075355 podStartE2EDuration="3.966075355s" podCreationTimestamp="2025-11-24 13:58:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:58:52.961195378 +0000 UTC m=+748.248952681" watchObservedRunningTime="2025-11-24 13:58:52.966075355 +0000 UTC m=+748.253832658" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.377348 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.377532 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-zd5kp" podUID="8f72328f-ea0c-4701-bd24-1e4cdf149962" containerName="registry-server" containerID="cri-o://3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8" gracePeriod=2 Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.832253 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.959146 4970 generic.go:334] "Generic (PLEG): container finished" podID="8f72328f-ea0c-4701-bd24-1e4cdf149962" containerID="3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8" exitCode=0 Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.959209 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zd5kp" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.959197 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zd5kp" event={"ID":"8f72328f-ea0c-4701-bd24-1e4cdf149962","Type":"ContainerDied","Data":"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8"} Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.959270 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zd5kp" event={"ID":"8f72328f-ea0c-4701-bd24-1e4cdf149962","Type":"ContainerDied","Data":"037c0fe1c593d85c1b51f85044413284ac34a3fd70d39ea26f1d2d216aa4459b"} Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.959290 4970 scope.go:117] "RemoveContainer" containerID="3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.963362 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7r4d\" (UniqueName: \"kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d\") pod \"8f72328f-ea0c-4701-bd24-1e4cdf149962\" (UID: \"8f72328f-ea0c-4701-bd24-1e4cdf149962\") " Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.970780 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d" (OuterVolumeSpecName: "kube-api-access-r7r4d") pod "8f72328f-ea0c-4701-bd24-1e4cdf149962" (UID: "8f72328f-ea0c-4701-bd24-1e4cdf149962"). InnerVolumeSpecName "kube-api-access-r7r4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.984060 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wrstd"] Nov 24 13:58:53 crc kubenswrapper[4970]: E1124 13:58:53.984324 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f72328f-ea0c-4701-bd24-1e4cdf149962" containerName="registry-server" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.984340 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f72328f-ea0c-4701-bd24-1e4cdf149962" containerName="registry-server" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.984476 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f72328f-ea0c-4701-bd24-1e4cdf149962" containerName="registry-server" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.991202 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.996684 4970 scope.go:117] "RemoveContainer" containerID="3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8" Nov 24 13:58:53 crc kubenswrapper[4970]: E1124 13:58:53.997857 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8\": container with ID starting with 3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8 not found: ID does not exist" containerID="3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.997982 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8"} err="failed to get container status \"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8\": rpc error: code = NotFound desc = could not find container \"3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8\": container with ID starting with 3d5a71ec1b4135ae28a7810ccf3b1e74d2464cc111535e649e022a9e7a8a3ae8 not found: ID does not exist" Nov 24 13:58:53 crc kubenswrapper[4970]: I1124 13:58:53.998495 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wrstd"] Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.066751 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-298f9\" (UniqueName: \"kubernetes.io/projected/2bca4967-78c0-444a-bd32-e12f72f64629-kube-api-access-298f9\") pod \"openstack-operator-index-wrstd\" (UID: \"2bca4967-78c0-444a-bd32-e12f72f64629\") " pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.067700 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7r4d\" (UniqueName: \"kubernetes.io/projected/8f72328f-ea0c-4701-bd24-1e4cdf149962-kube-api-access-r7r4d\") on node \"crc\" DevicePath \"\"" Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.169473 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-298f9\" (UniqueName: \"kubernetes.io/projected/2bca4967-78c0-444a-bd32-e12f72f64629-kube-api-access-298f9\") pod \"openstack-operator-index-wrstd\" (UID: \"2bca4967-78c0-444a-bd32-e12f72f64629\") " pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.190917 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-298f9\" (UniqueName: \"kubernetes.io/projected/2bca4967-78c0-444a-bd32-e12f72f64629-kube-api-access-298f9\") pod \"openstack-operator-index-wrstd\" (UID: \"2bca4967-78c0-444a-bd32-e12f72f64629\") " pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.303944 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.306995 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-zd5kp"] Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.327113 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.575543 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wrstd"] Nov 24 13:58:54 crc kubenswrapper[4970]: I1124 13:58:54.966853 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wrstd" event={"ID":"2bca4967-78c0-444a-bd32-e12f72f64629","Type":"ContainerStarted","Data":"89ccf0da365a3692673ab14acb94b8e6b6a5ff50d04dfafe7667f0f12c64805a"} Nov 24 13:58:55 crc kubenswrapper[4970]: I1124 13:58:55.480698 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f72328f-ea0c-4701-bd24-1e4cdf149962" path="/var/lib/kubelet/pods/8f72328f-ea0c-4701-bd24-1e4cdf149962/volumes" Nov 24 13:58:55 crc kubenswrapper[4970]: I1124 13:58:55.974593 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wrstd" event={"ID":"2bca4967-78c0-444a-bd32-e12f72f64629","Type":"ContainerStarted","Data":"75e2487da5c925489032fdfef170b3cdbf64b9f66a27ba9adcfe4e92a3718e27"} Nov 24 13:58:55 crc kubenswrapper[4970]: I1124 13:58:55.998733 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wrstd" podStartSLOduration=2.159673294 podStartE2EDuration="2.998706829s" podCreationTimestamp="2025-11-24 13:58:53 +0000 UTC" firstStartedPulling="2025-11-24 13:58:54.593412221 +0000 UTC m=+749.881169514" lastFinishedPulling="2025-11-24 13:58:55.432445756 +0000 UTC m=+750.720203049" observedRunningTime="2025-11-24 13:58:55.992463997 +0000 UTC m=+751.280221300" watchObservedRunningTime="2025-11-24 13:58:55.998706829 +0000 UTC m=+751.286464132" Nov 24 13:59:01 crc kubenswrapper[4970]: I1124 13:59:01.103449 4970 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 13:59:04 crc kubenswrapper[4970]: I1124 13:59:04.328950 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:59:04 crc kubenswrapper[4970]: I1124 13:59:04.330236 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:59:04 crc kubenswrapper[4970]: I1124 13:59:04.362865 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:59:05 crc kubenswrapper[4970]: I1124 13:59:05.059473 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-wrstd" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.213366 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr"] Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.216088 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.221300 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lxw4b" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.225964 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr"] Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.373967 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.374266 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.374317 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kppdq\" (UniqueName: \"kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.475948 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kppdq\" (UniqueName: \"kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.476453 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.477319 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.477555 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.478145 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.503142 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kppdq\" (UniqueName: \"kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq\") pod \"1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:06 crc kubenswrapper[4970]: I1124 13:59:06.584250 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:07 crc kubenswrapper[4970]: I1124 13:59:07.069367 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr"] Nov 24 13:59:08 crc kubenswrapper[4970]: I1124 13:59:08.056403 4970 generic.go:334] "Generic (PLEG): container finished" podID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerID="1a7c1cb7f3ac7f118922fadfcc62e239ad32d6b15f8097fd80c1c1393f17510c" exitCode=0 Nov 24 13:59:08 crc kubenswrapper[4970]: I1124 13:59:08.056471 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerDied","Data":"1a7c1cb7f3ac7f118922fadfcc62e239ad32d6b15f8097fd80c1c1393f17510c"} Nov 24 13:59:08 crc kubenswrapper[4970]: I1124 13:59:08.056510 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerStarted","Data":"82a452cc16e67571cf2b27daa6659722b95fd06a948909ac8f20a58d50947c63"} Nov 24 13:59:09 crc kubenswrapper[4970]: I1124 13:59:09.063474 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerStarted","Data":"1b1084c4741f3599afc3b957e50d13143830f3449daea302fa4c023b44ec7711"} Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.072719 4970 generic.go:334] "Generic (PLEG): container finished" podID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerID="1b1084c4741f3599afc3b957e50d13143830f3449daea302fa4c023b44ec7711" exitCode=0 Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.072863 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerDied","Data":"1b1084c4741f3599afc3b957e50d13143830f3449daea302fa4c023b44ec7711"} Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.606758 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.608966 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.625824 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.632857 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.632938 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.734178 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l998g\" (UniqueName: \"kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.734239 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.734265 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.734659 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.734741 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.835508 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l998g\" (UniqueName: \"kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.857501 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l998g\" (UniqueName: \"kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g\") pod \"redhat-operators-t5s8z\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:10 crc kubenswrapper[4970]: I1124 13:59:10.933027 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.082431 4970 generic.go:334] "Generic (PLEG): container finished" podID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerID="18967081ff83e1beea8c871c9e6302f7b6a735ab035741b0cecfa551168bfeb1" exitCode=0 Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.083067 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerDied","Data":"18967081ff83e1beea8c871c9e6302f7b6a735ab035741b0cecfa551168bfeb1"} Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.203934 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.203995 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.204034 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.204568 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.204636 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099" gracePeriod=600 Nov 24 13:59:11 crc kubenswrapper[4970]: I1124 13:59:11.394772 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 13:59:11 crc kubenswrapper[4970]: W1124 13:59:11.410902 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7015e74a_a3c2_4d9e_928b_ae37748bc3f8.slice/crio-bd10b4c54ff7d5b39db1f7df3397f7bce093948911acdd2b1bb99fdc15bf8263 WatchSource:0}: Error finding container bd10b4c54ff7d5b39db1f7df3397f7bce093948911acdd2b1bb99fdc15bf8263: Status 404 returned error can't find the container with id bd10b4c54ff7d5b39db1f7df3397f7bce093948911acdd2b1bb99fdc15bf8263 Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.098127 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099" exitCode=0 Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.098178 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099"} Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.098686 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c"} Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.098729 4970 scope.go:117] "RemoveContainer" containerID="f2501b6a3fcdda62b68f4c06654c813acb2b4c90cb7b2e758029e8d986f2c272" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.105132 4970 generic.go:334] "Generic (PLEG): container finished" podID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerID="91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4" exitCode=0 Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.105310 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerDied","Data":"91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4"} Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.105354 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerStarted","Data":"bd10b4c54ff7d5b39db1f7df3397f7bce093948911acdd2b1bb99fdc15bf8263"} Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.652767 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.760491 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle\") pod \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.760608 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util\") pod \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.760728 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kppdq\" (UniqueName: \"kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq\") pod \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\" (UID: \"373aeb29-00a5-4ee1-a007-fc9c113a99b6\") " Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.761489 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle" (OuterVolumeSpecName: "bundle") pod "373aeb29-00a5-4ee1-a007-fc9c113a99b6" (UID: "373aeb29-00a5-4ee1-a007-fc9c113a99b6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.766347 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq" (OuterVolumeSpecName: "kube-api-access-kppdq") pod "373aeb29-00a5-4ee1-a007-fc9c113a99b6" (UID: "373aeb29-00a5-4ee1-a007-fc9c113a99b6"). InnerVolumeSpecName "kube-api-access-kppdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.781375 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util" (OuterVolumeSpecName: "util") pod "373aeb29-00a5-4ee1-a007-fc9c113a99b6" (UID: "373aeb29-00a5-4ee1-a007-fc9c113a99b6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.863118 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kppdq\" (UniqueName: \"kubernetes.io/projected/373aeb29-00a5-4ee1-a007-fc9c113a99b6-kube-api-access-kppdq\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.863189 4970 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:12 crc kubenswrapper[4970]: I1124 13:59:12.863208 4970 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/373aeb29-00a5-4ee1-a007-fc9c113a99b6-util\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:13 crc kubenswrapper[4970]: I1124 13:59:13.114951 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" event={"ID":"373aeb29-00a5-4ee1-a007-fc9c113a99b6","Type":"ContainerDied","Data":"82a452cc16e67571cf2b27daa6659722b95fd06a948909ac8f20a58d50947c63"} Nov 24 13:59:13 crc kubenswrapper[4970]: I1124 13:59:13.114986 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82a452cc16e67571cf2b27daa6659722b95fd06a948909ac8f20a58d50947c63" Nov 24 13:59:13 crc kubenswrapper[4970]: I1124 13:59:13.114997 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.202455 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:15 crc kubenswrapper[4970]: E1124 13:59:15.203182 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="util" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.203204 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="util" Nov 24 13:59:15 crc kubenswrapper[4970]: E1124 13:59:15.203229 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="extract" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.203242 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="extract" Nov 24 13:59:15 crc kubenswrapper[4970]: E1124 13:59:15.203287 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="pull" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.203301 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="pull" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.203498 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="373aeb29-00a5-4ee1-a007-fc9c113a99b6" containerName="extract" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.205127 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.212283 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.304561 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.304663 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8599m\" (UniqueName: \"kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.304731 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.405323 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8599m\" (UniqueName: \"kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.405394 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.405465 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.406099 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.406342 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.425264 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8599m\" (UniqueName: \"kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m\") pod \"certified-operators-ncsrd\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:15 crc kubenswrapper[4970]: I1124 13:59:15.520243 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.044204 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.133247 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerStarted","Data":"fcf61c3f1e6df77221df3b88ac8030c4e84a4208742cb932bf164f2e3ba444ec"} Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.474692 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh"] Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.475491 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.477568 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-84fxq" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.503246 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh"] Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.624621 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw8xp\" (UniqueName: \"kubernetes.io/projected/f268a9ff-bab7-41a4-9d10-d0494d511f3a-kube-api-access-qw8xp\") pod \"openstack-operator-controller-operator-8486c7f98b-pb4vh\" (UID: \"f268a9ff-bab7-41a4-9d10-d0494d511f3a\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.726393 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw8xp\" (UniqueName: \"kubernetes.io/projected/f268a9ff-bab7-41a4-9d10-d0494d511f3a-kube-api-access-qw8xp\") pod \"openstack-operator-controller-operator-8486c7f98b-pb4vh\" (UID: \"f268a9ff-bab7-41a4-9d10-d0494d511f3a\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.755202 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw8xp\" (UniqueName: \"kubernetes.io/projected/f268a9ff-bab7-41a4-9d10-d0494d511f3a-kube-api-access-qw8xp\") pod \"openstack-operator-controller-operator-8486c7f98b-pb4vh\" (UID: \"f268a9ff-bab7-41a4-9d10-d0494d511f3a\") " pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:16 crc kubenswrapper[4970]: I1124 13:59:16.798549 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:17 crc kubenswrapper[4970]: I1124 13:59:17.140294 4970 generic.go:334] "Generic (PLEG): container finished" podID="85c09497-9643-40d0-a999-e20b376623e6" containerID="8f7f1897b2de6b1262a5621cb8670f1d946e05519cd91aa0a047434f3e07d0b0" exitCode=0 Nov 24 13:59:17 crc kubenswrapper[4970]: I1124 13:59:17.140335 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerDied","Data":"8f7f1897b2de6b1262a5621cb8670f1d946e05519cd91aa0a047434f3e07d0b0"} Nov 24 13:59:20 crc kubenswrapper[4970]: I1124 13:59:20.830968 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh"] Nov 24 13:59:21 crc kubenswrapper[4970]: I1124 13:59:21.167800 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerStarted","Data":"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f"} Nov 24 13:59:21 crc kubenswrapper[4970]: I1124 13:59:21.168930 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" event={"ID":"f268a9ff-bab7-41a4-9d10-d0494d511f3a","Type":"ContainerStarted","Data":"9e641920882b558492bfe27ef4963386deb078535ad400914e571715cbb42218"} Nov 24 13:59:22 crc kubenswrapper[4970]: I1124 13:59:22.179350 4970 generic.go:334] "Generic (PLEG): container finished" podID="85c09497-9643-40d0-a999-e20b376623e6" containerID="be65dcac206daddcae7405f8d5fd41930004eb5725ec43c038dadfc4a0cb9ee9" exitCode=0 Nov 24 13:59:22 crc kubenswrapper[4970]: I1124 13:59:22.179399 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerDied","Data":"be65dcac206daddcae7405f8d5fd41930004eb5725ec43c038dadfc4a0cb9ee9"} Nov 24 13:59:22 crc kubenswrapper[4970]: I1124 13:59:22.182828 4970 generic.go:334] "Generic (PLEG): container finished" podID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerID="2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f" exitCode=0 Nov 24 13:59:22 crc kubenswrapper[4970]: I1124 13:59:22.182884 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerDied","Data":"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f"} Nov 24 13:59:23 crc kubenswrapper[4970]: I1124 13:59:23.189766 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerStarted","Data":"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961"} Nov 24 13:59:23 crc kubenswrapper[4970]: I1124 13:59:23.204336 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerStarted","Data":"f09240e0d1cebb9361b626bdf6aa1eb1d47be2706578a6895c71b076e4ea29fa"} Nov 24 13:59:23 crc kubenswrapper[4970]: I1124 13:59:23.215262 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t5s8z" podStartSLOduration=2.754671618 podStartE2EDuration="13.215245417s" podCreationTimestamp="2025-11-24 13:59:10 +0000 UTC" firstStartedPulling="2025-11-24 13:59:12.108723421 +0000 UTC m=+767.396480754" lastFinishedPulling="2025-11-24 13:59:22.56929722 +0000 UTC m=+777.857054553" observedRunningTime="2025-11-24 13:59:23.209539559 +0000 UTC m=+778.497296862" watchObservedRunningTime="2025-11-24 13:59:23.215245417 +0000 UTC m=+778.503002710" Nov 24 13:59:23 crc kubenswrapper[4970]: I1124 13:59:23.229056 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ncsrd" podStartSLOduration=5.965584208 podStartE2EDuration="8.229036436s" podCreationTimestamp="2025-11-24 13:59:15 +0000 UTC" firstStartedPulling="2025-11-24 13:59:20.340889805 +0000 UTC m=+775.628647098" lastFinishedPulling="2025-11-24 13:59:22.604342033 +0000 UTC m=+777.892099326" observedRunningTime="2025-11-24 13:59:23.22725837 +0000 UTC m=+778.515015673" watchObservedRunningTime="2025-11-24 13:59:23.229036436 +0000 UTC m=+778.516793739" Nov 24 13:59:25 crc kubenswrapper[4970]: I1124 13:59:25.520884 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:25 crc kubenswrapper[4970]: I1124 13:59:25.521147 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:25 crc kubenswrapper[4970]: I1124 13:59:25.563404 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:26 crc kubenswrapper[4970]: I1124 13:59:26.236791 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" event={"ID":"f268a9ff-bab7-41a4-9d10-d0494d511f3a","Type":"ContainerStarted","Data":"4fab06bc2f8037dd9f000bf3438160d73e8482ffa631cbe6a356acdd046da40a"} Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.391860 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.394620 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.403223 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.498084 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg5hq\" (UniqueName: \"kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.498209 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.498243 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.599658 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg5hq\" (UniqueName: \"kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.599730 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.599752 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.727508 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.727654 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:27 crc kubenswrapper[4970]: I1124 13:59:27.728441 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg5hq\" (UniqueName: \"kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq\") pod \"community-operators-7nzql\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:28 crc kubenswrapper[4970]: I1124 13:59:28.017790 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:28 crc kubenswrapper[4970]: I1124 13:59:28.456957 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:28 crc kubenswrapper[4970]: W1124 13:59:28.462824 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd4cabde_b2b0_4ab4_a882_c237b7f6733a.slice/crio-1217d7cb6765b2767abe0cdb5d98519e0f76d7d20792fa970f8ba348135f725f WatchSource:0}: Error finding container 1217d7cb6765b2767abe0cdb5d98519e0f76d7d20792fa970f8ba348135f725f: Status 404 returned error can't find the container with id 1217d7cb6765b2767abe0cdb5d98519e0f76d7d20792fa970f8ba348135f725f Nov 24 13:59:29 crc kubenswrapper[4970]: I1124 13:59:29.257614 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerStarted","Data":"1217d7cb6765b2767abe0cdb5d98519e0f76d7d20792fa970f8ba348135f725f"} Nov 24 13:59:30 crc kubenswrapper[4970]: I1124 13:59:30.933968 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:30 crc kubenswrapper[4970]: I1124 13:59:30.934368 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:30 crc kubenswrapper[4970]: I1124 13:59:30.978215 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:31 crc kubenswrapper[4970]: I1124 13:59:31.295329 4970 generic.go:334] "Generic (PLEG): container finished" podID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerID="dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f" exitCode=0 Nov 24 13:59:31 crc kubenswrapper[4970]: I1124 13:59:31.296190 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerDied","Data":"dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f"} Nov 24 13:59:31 crc kubenswrapper[4970]: I1124 13:59:31.406830 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.046293 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.379524 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.380111 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frtlk" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="registry-server" containerID="cri-o://e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0" gracePeriod=2 Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.769519 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.810623 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content\") pod \"a73c8c7e-4407-4624-93cc-c13e6d56347a\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.810674 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdvnx\" (UniqueName: \"kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx\") pod \"a73c8c7e-4407-4624-93cc-c13e6d56347a\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.810812 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities\") pod \"a73c8c7e-4407-4624-93cc-c13e6d56347a\" (UID: \"a73c8c7e-4407-4624-93cc-c13e6d56347a\") " Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.811516 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities" (OuterVolumeSpecName: "utilities") pod "a73c8c7e-4407-4624-93cc-c13e6d56347a" (UID: "a73c8c7e-4407-4624-93cc-c13e6d56347a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.817669 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx" (OuterVolumeSpecName: "kube-api-access-vdvnx") pod "a73c8c7e-4407-4624-93cc-c13e6d56347a" (UID: "a73c8c7e-4407-4624-93cc-c13e6d56347a"). InnerVolumeSpecName "kube-api-access-vdvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.904969 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a73c8c7e-4407-4624-93cc-c13e6d56347a" (UID: "a73c8c7e-4407-4624-93cc-c13e6d56347a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.912699 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.912754 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c8c7e-4407-4624-93cc-c13e6d56347a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:33 crc kubenswrapper[4970]: I1124 13:59:33.912770 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdvnx\" (UniqueName: \"kubernetes.io/projected/a73c8c7e-4407-4624-93cc-c13e6d56347a-kube-api-access-vdvnx\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.343927 4970 generic.go:334] "Generic (PLEG): container finished" podID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerID="235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe" exitCode=0 Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.343980 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerDied","Data":"235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe"} Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.351382 4970 generic.go:334] "Generic (PLEG): container finished" podID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerID="e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0" exitCode=0 Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.351470 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frtlk" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.351479 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerDied","Data":"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0"} Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.351516 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frtlk" event={"ID":"a73c8c7e-4407-4624-93cc-c13e6d56347a","Type":"ContainerDied","Data":"187f4ecb5c66eadea0066bb269d82b2245a90b60c4b02179b3651571c9caf16f"} Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.351539 4970 scope.go:117] "RemoveContainer" containerID="e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.365085 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" event={"ID":"f268a9ff-bab7-41a4-9d10-d0494d511f3a","Type":"ContainerStarted","Data":"313803d950f2cd3b7c0ffa750d8947fa5ad5d7cd1923f639b77dca911c23bf48"} Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.365958 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.367682 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.375867 4970 scope.go:117] "RemoveContainer" containerID="c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.409614 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-8486c7f98b-pb4vh" podStartSLOduration=6.117491637 podStartE2EDuration="18.409594681s" podCreationTimestamp="2025-11-24 13:59:16 +0000 UTC" firstStartedPulling="2025-11-24 13:59:20.84623394 +0000 UTC m=+776.133991243" lastFinishedPulling="2025-11-24 13:59:33.138336954 +0000 UTC m=+788.426094287" observedRunningTime="2025-11-24 13:59:34.398743578 +0000 UTC m=+789.686500871" watchObservedRunningTime="2025-11-24 13:59:34.409594681 +0000 UTC m=+789.697351974" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.409666 4970 scope.go:117] "RemoveContainer" containerID="3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.411073 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.415787 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frtlk"] Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.430763 4970 scope.go:117] "RemoveContainer" containerID="e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0" Nov 24 13:59:34 crc kubenswrapper[4970]: E1124 13:59:34.433645 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0\": container with ID starting with e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0 not found: ID does not exist" containerID="e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.433676 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0"} err="failed to get container status \"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0\": rpc error: code = NotFound desc = could not find container \"e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0\": container with ID starting with e226d3389d68d17eb60db30a098d9029ad0f402aaca845587b7c9a23686554e0 not found: ID does not exist" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.433696 4970 scope.go:117] "RemoveContainer" containerID="c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1" Nov 24 13:59:34 crc kubenswrapper[4970]: E1124 13:59:34.434850 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1\": container with ID starting with c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1 not found: ID does not exist" containerID="c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.434965 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1"} err="failed to get container status \"c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1\": rpc error: code = NotFound desc = could not find container \"c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1\": container with ID starting with c0e2f1472594623ef3308692dca1520070fff735e3d58361d0e5be59a5a185b1 not found: ID does not exist" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.435041 4970 scope.go:117] "RemoveContainer" containerID="3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6" Nov 24 13:59:34 crc kubenswrapper[4970]: E1124 13:59:34.436179 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6\": container with ID starting with 3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6 not found: ID does not exist" containerID="3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6" Nov 24 13:59:34 crc kubenswrapper[4970]: I1124 13:59:34.436216 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6"} err="failed to get container status \"3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6\": rpc error: code = NotFound desc = could not find container \"3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6\": container with ID starting with 3a80fc5923013ae340a82074aa66d6ac9758d997c5abcd27c60ef1ed14d738a6 not found: ID does not exist" Nov 24 13:59:35 crc kubenswrapper[4970]: I1124 13:59:35.373061 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerStarted","Data":"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef"} Nov 24 13:59:35 crc kubenswrapper[4970]: I1124 13:59:35.399528 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7nzql" podStartSLOduration=5.281992729 podStartE2EDuration="8.399507123s" podCreationTimestamp="2025-11-24 13:59:27 +0000 UTC" firstStartedPulling="2025-11-24 13:59:31.647148421 +0000 UTC m=+786.934905754" lastFinishedPulling="2025-11-24 13:59:34.764662855 +0000 UTC m=+790.052420148" observedRunningTime="2025-11-24 13:59:35.394642586 +0000 UTC m=+790.682399879" watchObservedRunningTime="2025-11-24 13:59:35.399507123 +0000 UTC m=+790.687264416" Nov 24 13:59:35 crc kubenswrapper[4970]: I1124 13:59:35.477515 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" path="/var/lib/kubelet/pods/a73c8c7e-4407-4624-93cc-c13e6d56347a/volumes" Nov 24 13:59:35 crc kubenswrapper[4970]: I1124 13:59:35.581049 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:38 crc kubenswrapper[4970]: I1124 13:59:38.018349 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:38 crc kubenswrapper[4970]: I1124 13:59:38.018721 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:38 crc kubenswrapper[4970]: I1124 13:59:38.085970 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:38 crc kubenswrapper[4970]: I1124 13:59:38.779272 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:38 crc kubenswrapper[4970]: I1124 13:59:38.779544 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ncsrd" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="registry-server" containerID="cri-o://f09240e0d1cebb9361b626bdf6aa1eb1d47be2706578a6895c71b076e4ea29fa" gracePeriod=2 Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.405056 4970 generic.go:334] "Generic (PLEG): container finished" podID="85c09497-9643-40d0-a999-e20b376623e6" containerID="f09240e0d1cebb9361b626bdf6aa1eb1d47be2706578a6895c71b076e4ea29fa" exitCode=0 Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.405117 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerDied","Data":"f09240e0d1cebb9361b626bdf6aa1eb1d47be2706578a6895c71b076e4ea29fa"} Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.811344 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.907134 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content\") pod \"85c09497-9643-40d0-a999-e20b376623e6\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.907189 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities\") pod \"85c09497-9643-40d0-a999-e20b376623e6\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.907211 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8599m\" (UniqueName: \"kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m\") pod \"85c09497-9643-40d0-a999-e20b376623e6\" (UID: \"85c09497-9643-40d0-a999-e20b376623e6\") " Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.908853 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities" (OuterVolumeSpecName: "utilities") pod "85c09497-9643-40d0-a999-e20b376623e6" (UID: "85c09497-9643-40d0-a999-e20b376623e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.913945 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m" (OuterVolumeSpecName: "kube-api-access-8599m") pod "85c09497-9643-40d0-a999-e20b376623e6" (UID: "85c09497-9643-40d0-a999-e20b376623e6"). InnerVolumeSpecName "kube-api-access-8599m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:59:39 crc kubenswrapper[4970]: I1124 13:59:39.988291 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85c09497-9643-40d0-a999-e20b376623e6" (UID: "85c09497-9643-40d0-a999-e20b376623e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.009377 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.009422 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c09497-9643-40d0-a999-e20b376623e6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.009438 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8599m\" (UniqueName: \"kubernetes.io/projected/85c09497-9643-40d0-a999-e20b376623e6-kube-api-access-8599m\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.412502 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ncsrd" event={"ID":"85c09497-9643-40d0-a999-e20b376623e6","Type":"ContainerDied","Data":"fcf61c3f1e6df77221df3b88ac8030c4e84a4208742cb932bf164f2e3ba444ec"} Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.412550 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ncsrd" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.412590 4970 scope.go:117] "RemoveContainer" containerID="f09240e0d1cebb9361b626bdf6aa1eb1d47be2706578a6895c71b076e4ea29fa" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.428752 4970 scope.go:117] "RemoveContainer" containerID="be65dcac206daddcae7405f8d5fd41930004eb5725ec43c038dadfc4a0cb9ee9" Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.444488 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.450354 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ncsrd"] Nov 24 13:59:40 crc kubenswrapper[4970]: I1124 13:59:40.461733 4970 scope.go:117] "RemoveContainer" containerID="8f7f1897b2de6b1262a5621cb8670f1d946e05519cd91aa0a047434f3e07d0b0" Nov 24 13:59:41 crc kubenswrapper[4970]: I1124 13:59:41.478355 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c09497-9643-40d0-a999-e20b376623e6" path="/var/lib/kubelet/pods/85c09497-9643-40d0-a999-e20b376623e6/volumes" Nov 24 13:59:48 crc kubenswrapper[4970]: I1124 13:59:48.064786 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:48 crc kubenswrapper[4970]: I1124 13:59:48.117560 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:48 crc kubenswrapper[4970]: I1124 13:59:48.466872 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7nzql" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="registry-server" containerID="cri-o://a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef" gracePeriod=2 Nov 24 13:59:48 crc kubenswrapper[4970]: I1124 13:59:48.948362 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.036213 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg5hq\" (UniqueName: \"kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq\") pod \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.036345 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content\") pod \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.036389 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities\") pod \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\" (UID: \"dd4cabde-b2b0-4ab4-a882-c237b7f6733a\") " Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.037122 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities" (OuterVolumeSpecName: "utilities") pod "dd4cabde-b2b0-4ab4-a882-c237b7f6733a" (UID: "dd4cabde-b2b0-4ab4-a882-c237b7f6733a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.044731 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq" (OuterVolumeSpecName: "kube-api-access-kg5hq") pod "dd4cabde-b2b0-4ab4-a882-c237b7f6733a" (UID: "dd4cabde-b2b0-4ab4-a882-c237b7f6733a"). InnerVolumeSpecName "kube-api-access-kg5hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.085233 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd4cabde-b2b0-4ab4-a882-c237b7f6733a" (UID: "dd4cabde-b2b0-4ab4-a882-c237b7f6733a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.137940 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.137971 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.137980 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg5hq\" (UniqueName: \"kubernetes.io/projected/dd4cabde-b2b0-4ab4-a882-c237b7f6733a-kube-api-access-kg5hq\") on node \"crc\" DevicePath \"\"" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.475691 4970 generic.go:334] "Generic (PLEG): container finished" podID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerID="a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef" exitCode=0 Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.475808 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7nzql" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.478648 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerDied","Data":"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef"} Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.478695 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7nzql" event={"ID":"dd4cabde-b2b0-4ab4-a882-c237b7f6733a","Type":"ContainerDied","Data":"1217d7cb6765b2767abe0cdb5d98519e0f76d7d20792fa970f8ba348135f725f"} Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.478718 4970 scope.go:117] "RemoveContainer" containerID="a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.494901 4970 scope.go:117] "RemoveContainer" containerID="235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.511060 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.520412 4970 scope.go:117] "RemoveContainer" containerID="dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.521513 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7nzql"] Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.563460 4970 scope.go:117] "RemoveContainer" containerID="a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef" Nov 24 13:59:49 crc kubenswrapper[4970]: E1124 13:59:49.564971 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef\": container with ID starting with a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef not found: ID does not exist" containerID="a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.565003 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef"} err="failed to get container status \"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef\": rpc error: code = NotFound desc = could not find container \"a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef\": container with ID starting with a187f9f23a2c939868236d88380b21618325b95ec535d1bd205af7ffaac848ef not found: ID does not exist" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.565024 4970 scope.go:117] "RemoveContainer" containerID="235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe" Nov 24 13:59:49 crc kubenswrapper[4970]: E1124 13:59:49.567938 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe\": container with ID starting with 235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe not found: ID does not exist" containerID="235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.567970 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe"} err="failed to get container status \"235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe\": rpc error: code = NotFound desc = could not find container \"235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe\": container with ID starting with 235c55ce453e0b4ffb1f6bdba3dd9b60d37c622fa5215846380e1e1addd96cbe not found: ID does not exist" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.567989 4970 scope.go:117] "RemoveContainer" containerID="dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f" Nov 24 13:59:49 crc kubenswrapper[4970]: E1124 13:59:49.568756 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f\": container with ID starting with dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f not found: ID does not exist" containerID="dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f" Nov 24 13:59:49 crc kubenswrapper[4970]: I1124 13:59:49.568780 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f"} err="failed to get container status \"dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f\": rpc error: code = NotFound desc = could not find container \"dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f\": container with ID starting with dcc7980378f4b0afd167d32a7b17a5622a547fe68853d571bcf9ddc0465e578f not found: ID does not exist" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.482256 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" path="/var/lib/kubelet/pods/dd4cabde-b2b0-4ab4-a882-c237b7f6733a/volumes" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.854748 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m"] Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855034 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855060 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855079 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855088 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855102 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855110 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855125 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855135 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855144 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855152 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855161 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855169 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="extract-utilities" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855187 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855195 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855205 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855212 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: E1124 13:59:51.855223 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855231 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="extract-content" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855371 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c09497-9643-40d0-a999-e20b376623e6" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855394 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73c8c7e-4407-4624-93cc-c13e6d56347a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.855406 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd4cabde-b2b0-4ab4-a882-c237b7f6733a" containerName="registry-server" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.856150 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.868095 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m"] Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.882160 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wvd5q" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.899983 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l"] Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.916865 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.939061 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8qj4h" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.949293 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c"] Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.950286 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.957795 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-5wgxh" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.960919 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx"] Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.962162 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.964232 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-wbwrr" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.977874 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm8gb\" (UniqueName: \"kubernetes.io/projected/8874383f-46c3-43b7-b1d1-56c9d14a0004-kube-api-access-lm8gb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-r452c\" (UID: \"8874383f-46c3-43b7-b1d1-56c9d14a0004\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.977917 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsjzs\" (UniqueName: \"kubernetes.io/projected/93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5-kube-api-access-rsjzs\") pod \"barbican-operator-controller-manager-7768f8c84f-kdx5m\" (UID: \"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.977981 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrmd9\" (UniqueName: \"kubernetes.io/projected/799af0bf-24f0-4546-960b-2e8e72083fb8-kube-api-access-nrmd9\") pod \"designate-operator-controller-manager-56dfb6b67f-ff65l\" (UID: \"799af0bf-24f0-4546-960b-2e8e72083fb8\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 13:59:51 crc kubenswrapper[4970]: I1124 13:59:51.995841 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.002484 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.022366 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.023570 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.035026 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-v5d4r" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.035177 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.047175 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.048102 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.051496 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.053044 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-cv8rc" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.069094 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079458 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz2mh\" (UniqueName: \"kubernetes.io/projected/9ca1ba14-aafe-429b-be3d-87dadeb73b54-kube-api-access-jz2mh\") pod \"heat-operator-controller-manager-bf4c6585d-v67fx\" (UID: \"9ca1ba14-aafe-429b-be3d-87dadeb73b54\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079622 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrmd9\" (UniqueName: \"kubernetes.io/projected/799af0bf-24f0-4546-960b-2e8e72083fb8-kube-api-access-nrmd9\") pod \"designate-operator-controller-manager-56dfb6b67f-ff65l\" (UID: \"799af0bf-24f0-4546-960b-2e8e72083fb8\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079728 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptgv4\" (UniqueName: \"kubernetes.io/projected/4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd-kube-api-access-ptgv4\") pod \"horizon-operator-controller-manager-5d86b44686-crqqj\" (UID: \"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079785 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm8gb\" (UniqueName: \"kubernetes.io/projected/8874383f-46c3-43b7-b1d1-56c9d14a0004-kube-api-access-lm8gb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-r452c\" (UID: \"8874383f-46c3-43b7-b1d1-56c9d14a0004\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079803 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsjzs\" (UniqueName: \"kubernetes.io/projected/93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5-kube-api-access-rsjzs\") pod \"barbican-operator-controller-manager-7768f8c84f-kdx5m\" (UID: \"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.079850 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl72f\" (UniqueName: \"kubernetes.io/projected/a5f4428d-ecd0-4c71-b9da-de188497a191-kube-api-access-hl72f\") pod \"glance-operator-controller-manager-8667fbf6f6-qclm2\" (UID: \"a5f4428d-ecd0-4c71-b9da-de188497a191\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.087677 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.088642 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.093886 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.094176 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-dz8p5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.103541 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.117766 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsjzs\" (UniqueName: \"kubernetes.io/projected/93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5-kube-api-access-rsjzs\") pod \"barbican-operator-controller-manager-7768f8c84f-kdx5m\" (UID: \"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5\") " pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.126220 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm8gb\" (UniqueName: \"kubernetes.io/projected/8874383f-46c3-43b7-b1d1-56c9d14a0004-kube-api-access-lm8gb\") pod \"cinder-operator-controller-manager-6d8fd67bf7-r452c\" (UID: \"8874383f-46c3-43b7-b1d1-56c9d14a0004\") " pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.143761 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrmd9\" (UniqueName: \"kubernetes.io/projected/799af0bf-24f0-4546-960b-2e8e72083fb8-kube-api-access-nrmd9\") pod \"designate-operator-controller-manager-56dfb6b67f-ff65l\" (UID: \"799af0bf-24f0-4546-960b-2e8e72083fb8\") " pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.145852 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.147103 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.151919 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-hfrw8" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.167942 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.170348 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.171287 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.177172 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.178112 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.178837 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rzs59" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180686 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2g4n\" (UniqueName: \"kubernetes.io/projected/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-kube-api-access-v2g4n\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180728 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shxcx\" (UniqueName: \"kubernetes.io/projected/f01cc05b-92d6-4144-ad87-5fdec63cc95c-kube-api-access-shxcx\") pod \"ironic-operator-controller-manager-5c75d7c94b-sx2v5\" (UID: \"f01cc05b-92d6-4144-ad87-5fdec63cc95c\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180758 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz2mh\" (UniqueName: \"kubernetes.io/projected/9ca1ba14-aafe-429b-be3d-87dadeb73b54-kube-api-access-jz2mh\") pod \"heat-operator-controller-manager-bf4c6585d-v67fx\" (UID: \"9ca1ba14-aafe-429b-be3d-87dadeb73b54\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180789 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bz9z\" (UniqueName: \"kubernetes.io/projected/73f36b2f-3ac8-4d5c-859b-1e376b9a5194-kube-api-access-5bz9z\") pod \"manila-operator-controller-manager-7bb88cb858-h9j24\" (UID: \"73f36b2f-3ac8-4d5c-859b-1e376b9a5194\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180840 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptgv4\" (UniqueName: \"kubernetes.io/projected/4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd-kube-api-access-ptgv4\") pod \"horizon-operator-controller-manager-5d86b44686-crqqj\" (UID: \"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180856 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-cert\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180899 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl72f\" (UniqueName: \"kubernetes.io/projected/a5f4428d-ecd0-4c71-b9da-de188497a191-kube-api-access-hl72f\") pod \"glance-operator-controller-manager-8667fbf6f6-qclm2\" (UID: \"a5f4428d-ecd0-4c71-b9da-de188497a191\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.180983 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bvsj5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.198743 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.202176 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.203868 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptgv4\" (UniqueName: \"kubernetes.io/projected/4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd-kube-api-access-ptgv4\") pod \"horizon-operator-controller-manager-5d86b44686-crqqj\" (UID: \"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd\") " pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.204995 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl72f\" (UniqueName: \"kubernetes.io/projected/a5f4428d-ecd0-4c71-b9da-de188497a191-kube-api-access-hl72f\") pod \"glance-operator-controller-manager-8667fbf6f6-qclm2\" (UID: \"a5f4428d-ecd0-4c71-b9da-de188497a191\") " pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.205259 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz2mh\" (UniqueName: \"kubernetes.io/projected/9ca1ba14-aafe-429b-be3d-87dadeb73b54-kube-api-access-jz2mh\") pod \"heat-operator-controller-manager-bf4c6585d-v67fx\" (UID: \"9ca1ba14-aafe-429b-be3d-87dadeb73b54\") " pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.207743 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.214875 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.215805 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.218840 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-6zxnq" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.229543 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.239618 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.240604 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.245862 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9ftqp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.248198 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-vs899"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.249202 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.250677 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bvp8s" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.253817 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.254866 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.260636 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8k2lv" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.265938 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.271502 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-vs899"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282229 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2g4n\" (UniqueName: \"kubernetes.io/projected/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-kube-api-access-v2g4n\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282270 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shxcx\" (UniqueName: \"kubernetes.io/projected/f01cc05b-92d6-4144-ad87-5fdec63cc95c-kube-api-access-shxcx\") pod \"ironic-operator-controller-manager-5c75d7c94b-sx2v5\" (UID: \"f01cc05b-92d6-4144-ad87-5fdec63cc95c\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282291 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9db6\" (UniqueName: \"kubernetes.io/projected/20a43498-16e1-46d8-9487-b87ff9cad8f8-kube-api-access-j9db6\") pod \"keystone-operator-controller-manager-7879fb76fd-9mnkp\" (UID: \"20a43498-16e1-46d8-9487-b87ff9cad8f8\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282312 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2dw\" (UniqueName: \"kubernetes.io/projected/b070979e-cf9f-4793-a50a-f0e67a48a58f-kube-api-access-6d2dw\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-wfsc5\" (UID: \"b070979e-cf9f-4793-a50a-f0e67a48a58f\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282341 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bz9z\" (UniqueName: \"kubernetes.io/projected/73f36b2f-3ac8-4d5c-859b-1e376b9a5194-kube-api-access-5bz9z\") pod \"manila-operator-controller-manager-7bb88cb858-h9j24\" (UID: \"73f36b2f-3ac8-4d5c-859b-1e376b9a5194\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282362 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7qx\" (UniqueName: \"kubernetes.io/projected/d94c9e13-b160-4f65-a97c-9bfdd4799eb0-kube-api-access-pl7qx\") pod \"octavia-operator-controller-manager-6fdc856c5d-g8kmd\" (UID: \"d94c9e13-b160-4f65-a97c-9bfdd4799eb0\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282397 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfx24\" (UniqueName: \"kubernetes.io/projected/a44bf58b-6eb3-4bdc-ad1c-41aa7623666b-kube-api-access-bfx24\") pod \"nova-operator-controller-manager-86d796d84d-vs899\" (UID: \"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282421 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-cert\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282552 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282881 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvnpn\" (UniqueName: \"kubernetes.io/projected/028f94a9-1db9-46e8-93f1-106d959ebe9b-kube-api-access-pvnpn\") pod \"neutron-operator-controller-manager-66b7d6f598-zkcwk\" (UID: \"028f94a9-1db9-46e8-93f1-106d959ebe9b\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.282916 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.295488 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.296468 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.298712 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-cert\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.298891 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-6lkmt" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.301100 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bz9z\" (UniqueName: \"kubernetes.io/projected/73f36b2f-3ac8-4d5c-859b-1e376b9a5194-kube-api-access-5bz9z\") pod \"manila-operator-controller-manager-7bb88cb858-h9j24\" (UID: \"73f36b2f-3ac8-4d5c-859b-1e376b9a5194\") " pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.302271 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2g4n\" (UniqueName: \"kubernetes.io/projected/d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0-kube-api-access-v2g4n\") pod \"infra-operator-controller-manager-769d9c7585-9lhqx\" (UID: \"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0\") " pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.302317 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.303262 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.303922 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shxcx\" (UniqueName: \"kubernetes.io/projected/f01cc05b-92d6-4144-ad87-5fdec63cc95c-kube-api-access-shxcx\") pod \"ironic-operator-controller-manager-5c75d7c94b-sx2v5\" (UID: \"f01cc05b-92d6-4144-ad87-5fdec63cc95c\") " pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.306009 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-x55lr" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.307537 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.307852 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.309354 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.312377 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.319979 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.338017 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.338275 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fwfk5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.339639 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.355081 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385513 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7dvl\" (UniqueName: \"kubernetes.io/projected/90e1030d-d9d0-42bf-b35b-a2f19ee0e6de-kube-api-access-t7dvl\") pod \"placement-operator-controller-manager-6dc664666c-t4pqg\" (UID: \"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385568 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd5g7\" (UniqueName: \"kubernetes.io/projected/d17a257a-f687-4d53-9b81-5fa430e24d70-kube-api-access-zd5g7\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385613 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9db6\" (UniqueName: \"kubernetes.io/projected/20a43498-16e1-46d8-9487-b87ff9cad8f8-kube-api-access-j9db6\") pod \"keystone-operator-controller-manager-7879fb76fd-9mnkp\" (UID: \"20a43498-16e1-46d8-9487-b87ff9cad8f8\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385641 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2dw\" (UniqueName: \"kubernetes.io/projected/b070979e-cf9f-4793-a50a-f0e67a48a58f-kube-api-access-6d2dw\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-wfsc5\" (UID: \"b070979e-cf9f-4793-a50a-f0e67a48a58f\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385676 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7qx\" (UniqueName: \"kubernetes.io/projected/d94c9e13-b160-4f65-a97c-9bfdd4799eb0-kube-api-access-pl7qx\") pod \"octavia-operator-controller-manager-6fdc856c5d-g8kmd\" (UID: \"d94c9e13-b160-4f65-a97c-9bfdd4799eb0\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385713 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385740 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfx24\" (UniqueName: \"kubernetes.io/projected/a44bf58b-6eb3-4bdc-ad1c-41aa7623666b-kube-api-access-bfx24\") pod \"nova-operator-controller-manager-86d796d84d-vs899\" (UID: \"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385776 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvnpn\" (UniqueName: \"kubernetes.io/projected/028f94a9-1db9-46e8-93f1-106d959ebe9b-kube-api-access-pvnpn\") pod \"neutron-operator-controller-manager-66b7d6f598-zkcwk\" (UID: \"028f94a9-1db9-46e8-93f1-106d959ebe9b\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385799 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76hbt\" (UniqueName: \"kubernetes.io/projected/1af43cf6-5835-47eb-a315-c16bf5758a9e-kube-api-access-76hbt\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-hdflg\" (UID: \"1af43cf6-5835-47eb-a315-c16bf5758a9e\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.385992 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.409167 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.421705 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfx24\" (UniqueName: \"kubernetes.io/projected/a44bf58b-6eb3-4bdc-ad1c-41aa7623666b-kube-api-access-bfx24\") pod \"nova-operator-controller-manager-86d796d84d-vs899\" (UID: \"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b\") " pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.425013 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvnpn\" (UniqueName: \"kubernetes.io/projected/028f94a9-1db9-46e8-93f1-106d959ebe9b-kube-api-access-pvnpn\") pod \"neutron-operator-controller-manager-66b7d6f598-zkcwk\" (UID: \"028f94a9-1db9-46e8-93f1-106d959ebe9b\") " pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.429236 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7qx\" (UniqueName: \"kubernetes.io/projected/d94c9e13-b160-4f65-a97c-9bfdd4799eb0-kube-api-access-pl7qx\") pod \"octavia-operator-controller-manager-6fdc856c5d-g8kmd\" (UID: \"d94c9e13-b160-4f65-a97c-9bfdd4799eb0\") " pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.449977 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9db6\" (UniqueName: \"kubernetes.io/projected/20a43498-16e1-46d8-9487-b87ff9cad8f8-kube-api-access-j9db6\") pod \"keystone-operator-controller-manager-7879fb76fd-9mnkp\" (UID: \"20a43498-16e1-46d8-9487-b87ff9cad8f8\") " pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.450566 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2dw\" (UniqueName: \"kubernetes.io/projected/b070979e-cf9f-4793-a50a-f0e67a48a58f-kube-api-access-6d2dw\") pod \"mariadb-operator-controller-manager-6f8c5b86cb-wfsc5\" (UID: \"b070979e-cf9f-4793-a50a-f0e67a48a58f\") " pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.497644 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.498722 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.499258 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.506984 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-ttzgp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.509204 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd5g7\" (UniqueName: \"kubernetes.io/projected/d17a257a-f687-4d53-9b81-5fa430e24d70-kube-api-access-zd5g7\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.509288 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.509343 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76hbt\" (UniqueName: \"kubernetes.io/projected/1af43cf6-5835-47eb-a315-c16bf5758a9e-kube-api-access-76hbt\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-hdflg\" (UID: \"1af43cf6-5835-47eb-a315-c16bf5758a9e\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.509384 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7dvl\" (UniqueName: \"kubernetes.io/projected/90e1030d-d9d0-42bf-b35b-a2f19ee0e6de-kube-api-access-t7dvl\") pod \"placement-operator-controller-manager-6dc664666c-t4pqg\" (UID: \"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 13:59:52 crc kubenswrapper[4970]: E1124 13:59:52.509786 4970 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:59:52 crc kubenswrapper[4970]: E1124 13:59:52.509832 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert podName:d17a257a-f687-4d53-9b81-5fa430e24d70 nodeName:}" failed. No retries permitted until 2025-11-24 13:59:53.009816708 +0000 UTC m=+808.297574001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert") pod "openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" (UID: "d17a257a-f687-4d53-9b81-5fa430e24d70") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.547831 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.559842 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.564237 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7dvl\" (UniqueName: \"kubernetes.io/projected/90e1030d-d9d0-42bf-b35b-a2f19ee0e6de-kube-api-access-t7dvl\") pod \"placement-operator-controller-manager-6dc664666c-t4pqg\" (UID: \"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de\") " pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.565080 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76hbt\" (UniqueName: \"kubernetes.io/projected/1af43cf6-5835-47eb-a315-c16bf5758a9e-kube-api-access-76hbt\") pod \"ovn-operator-controller-manager-5bdf4f7f7f-hdflg\" (UID: \"1af43cf6-5835-47eb-a315-c16bf5758a9e\") " pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.568977 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.576659 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd5g7\" (UniqueName: \"kubernetes.io/projected/d17a257a-f687-4d53-9b81-5fa430e24d70-kube-api-access-zd5g7\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.588203 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.601460 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.602564 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.605190 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.615334 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-rvr6h" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.617514 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crk2n\" (UniqueName: \"kubernetes.io/projected/0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20-kube-api-access-crk2n\") pod \"swift-operator-controller-manager-799cb6ffd6-5sp97\" (UID: \"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.617621 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrrr\" (UniqueName: \"kubernetes.io/projected/289cc484-057a-4107-937d-6c532c21b52a-kube-api-access-slrrr\") pod \"telemetry-operator-controller-manager-7798859c74-k5b5j\" (UID: \"289cc484-057a-4107-937d-6c532c21b52a\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.648684 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.680499 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.699125 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.714614 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.719165 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crk2n\" (UniqueName: \"kubernetes.io/projected/0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20-kube-api-access-crk2n\") pod \"swift-operator-controller-manager-799cb6ffd6-5sp97\" (UID: \"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.719490 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrrr\" (UniqueName: \"kubernetes.io/projected/289cc484-057a-4107-937d-6c532c21b52a-kube-api-access-slrrr\") pod \"telemetry-operator-controller-manager-7798859c74-k5b5j\" (UID: \"289cc484-057a-4107-937d-6c532c21b52a\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.723732 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.742972 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrrr\" (UniqueName: \"kubernetes.io/projected/289cc484-057a-4107-937d-6c532c21b52a-kube-api-access-slrrr\") pod \"telemetry-operator-controller-manager-7798859c74-k5b5j\" (UID: \"289cc484-057a-4107-937d-6c532c21b52a\") " pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.743460 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.747234 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crk2n\" (UniqueName: \"kubernetes.io/projected/0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20-kube-api-access-crk2n\") pod \"swift-operator-controller-manager-799cb6ffd6-5sp97\" (UID: \"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20\") " pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.762736 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.763881 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.770742 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-b5k6x" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.779043 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.820169 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpj5m\" (UniqueName: \"kubernetes.io/projected/5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7-kube-api-access-vpj5m\") pod \"test-operator-controller-manager-8464cf66df-wfmzk\" (UID: \"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.828045 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.829123 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.829315 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.834501 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-642g9" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.880864 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.894522 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.895965 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.898656 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-j4mwh" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.898836 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.899081 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.903189 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.911938 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.912779 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd"] Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.912872 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.915349 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-dlkbk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.922925 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpj5m\" (UniqueName: \"kubernetes.io/projected/5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7-kube-api-access-vpj5m\") pod \"test-operator-controller-manager-8464cf66df-wfmzk\" (UID: \"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.952628 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpj5m\" (UniqueName: \"kubernetes.io/projected/5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7-kube-api-access-vpj5m\") pod \"test-operator-controller-manager-8464cf66df-wfmzk\" (UID: \"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7\") " pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 13:59:52 crc kubenswrapper[4970]: I1124 13:59:52.961229 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.024985 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd2wh\" (UniqueName: \"kubernetes.io/projected/9847f51b-ce73-493a-ac47-b93341cebf73-kube-api-access-bd2wh\") pod \"watcher-operator-controller-manager-7cd4fb6f79-m9m2l\" (UID: \"9847f51b-ce73-493a-ac47-b93341cebf73\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.025036 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z5tm\" (UniqueName: \"kubernetes.io/projected/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-kube-api-access-8z5tm\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.025067 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.025097 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cphvx\" (UniqueName: \"kubernetes.io/projected/65008609-c137-4dee-b406-159814902ff4-kube-api-access-cphvx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-488cd\" (UID: \"65008609-c137-4dee-b406-159814902ff4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.025116 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.028552 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d17a257a-f687-4d53-9b81-5fa430e24d70-cert\") pod \"openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp\" (UID: \"d17a257a-f687-4d53-9b81-5fa430e24d70\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.109537 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.122283 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.126481 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cphvx\" (UniqueName: \"kubernetes.io/projected/65008609-c137-4dee-b406-159814902ff4-kube-api-access-cphvx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-488cd\" (UID: \"65008609-c137-4dee-b406-159814902ff4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.126530 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.126638 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd2wh\" (UniqueName: \"kubernetes.io/projected/9847f51b-ce73-493a-ac47-b93341cebf73-kube-api-access-bd2wh\") pod \"watcher-operator-controller-manager-7cd4fb6f79-m9m2l\" (UID: \"9847f51b-ce73-493a-ac47-b93341cebf73\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.126685 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z5tm\" (UniqueName: \"kubernetes.io/projected/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-kube-api-access-8z5tm\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: E1124 13:59:53.126953 4970 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 13:59:53 crc kubenswrapper[4970]: E1124 13:59:53.127028 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert podName:6ed5751d-33ce-4775-a0c7-cb951a6fbdcb nodeName:}" failed. No retries permitted until 2025-11-24 13:59:53.627008857 +0000 UTC m=+808.914766150 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert") pod "openstack-operator-controller-manager-6cb9dc54f8-bcbxx" (UID: "6ed5751d-33ce-4775-a0c7-cb951a6fbdcb") : secret "webhook-server-cert" not found Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.151312 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z5tm\" (UniqueName: \"kubernetes.io/projected/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-kube-api-access-8z5tm\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.152866 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.154325 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cphvx\" (UniqueName: \"kubernetes.io/projected/65008609-c137-4dee-b406-159814902ff4-kube-api-access-cphvx\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-488cd\" (UID: \"65008609-c137-4dee-b406-159814902ff4\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.156795 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd2wh\" (UniqueName: \"kubernetes.io/projected/9847f51b-ce73-493a-ac47-b93341cebf73-kube-api-access-bd2wh\") pod \"watcher-operator-controller-manager-7cd4fb6f79-m9m2l\" (UID: \"9847f51b-ce73-493a-ac47-b93341cebf73\") " pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.162366 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.172728 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 13:59:53 crc kubenswrapper[4970]: W1124 13:59:53.262148 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8874383f_46c3_43b7_b1d1_56c9d14a0004.slice/crio-3f27d1e2e9c54cbc9f1b634ad6e48e9f94944fa48673b4a9ccbf560084640c99 WatchSource:0}: Error finding container 3f27d1e2e9c54cbc9f1b634ad6e48e9f94944fa48673b4a9ccbf560084640c99: Status 404 returned error can't find the container with id 3f27d1e2e9c54cbc9f1b634ad6e48e9f94944fa48673b4a9ccbf560084640c99 Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.291067 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.323975 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.374955 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.529742 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.546974 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.550811 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" event={"ID":"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5","Type":"ContainerStarted","Data":"371ed42d8f3d24ee664a797ec9d98cfd01a8024e7ce62061601df6e9662d7a79"} Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.552245 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" event={"ID":"9ca1ba14-aafe-429b-be3d-87dadeb73b54","Type":"ContainerStarted","Data":"6d6ca7fbeacf8228e51ec993e3fb48e17660dc71f79724bd3e9c327d95d5eb79"} Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.553842 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" event={"ID":"8874383f-46c3-43b7-b1d1-56c9d14a0004","Type":"ContainerStarted","Data":"3f27d1e2e9c54cbc9f1b634ad6e48e9f94944fa48673b4a9ccbf560084640c99"} Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.560207 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" event={"ID":"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0","Type":"ContainerStarted","Data":"21c81b68ac1b729e8827361e73d8c4e72c20a2a46a9f7299226489bac913e170"} Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.560259 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.562218 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" event={"ID":"799af0bf-24f0-4546-960b-2e8e72083fb8","Type":"ContainerStarted","Data":"9a7d9448a1400142dbaac5624cc20e66ac93311fb4c77785408eb50f54d03659"} Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.564822 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" event={"ID":"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd","Type":"ContainerStarted","Data":"14747fa3d1144e2c1e98dc634154a6dcf765657bd0997a856a86fc479724e3c6"} Nov 24 13:59:53 crc kubenswrapper[4970]: W1124 13:59:53.566312 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf01cc05b_92d6_4144_ad87_5fdec63cc95c.slice/crio-237624fed825cd1fe12c247b8cfb5222d309ad2ac06ffb9283c02a4bc74ae248 WatchSource:0}: Error finding container 237624fed825cd1fe12c247b8cfb5222d309ad2ac06ffb9283c02a4bc74ae248: Status 404 returned error can't find the container with id 237624fed825cd1fe12c247b8cfb5222d309ad2ac06ffb9283c02a4bc74ae248 Nov 24 13:59:53 crc kubenswrapper[4970]: W1124 13:59:53.568004 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73f36b2f_3ac8_4d5c_859b_1e376b9a5194.slice/crio-63d4b57a52497888e13651956e3d77baa5bbc21f7fb82962f8a74321e27be94d WatchSource:0}: Error finding container 63d4b57a52497888e13651956e3d77baa5bbc21f7fb82962f8a74321e27be94d: Status 404 returned error can't find the container with id 63d4b57a52497888e13651956e3d77baa5bbc21f7fb82962f8a74321e27be94d Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.589651 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2"] Nov 24 13:59:53 crc kubenswrapper[4970]: W1124 13:59:53.595373 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5f4428d_ecd0_4c71_b9da_de188497a191.slice/crio-998c3d1ee574e47f94847ca4c8fca3eb109e6be5097281add8e8ce8cfae35519 WatchSource:0}: Error finding container 998c3d1ee574e47f94847ca4c8fca3eb109e6be5097281add8e8ce8cfae35519: Status 404 returned error can't find the container with id 998c3d1ee574e47f94847ca4c8fca3eb109e6be5097281add8e8ce8cfae35519 Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.634661 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.640650 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ed5751d-33ce-4775-a0c7-cb951a6fbdcb-cert\") pod \"openstack-operator-controller-manager-6cb9dc54f8-bcbxx\" (UID: \"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb\") " pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.874882 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.972667 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.982711 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk"] Nov 24 13:59:53 crc kubenswrapper[4970]: I1124 13:59:53.994157 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk"] Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.007122 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg"] Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.012300 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg"] Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.023598 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5"] Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.029176 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t7dvl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-6dc664666c-t4pqg_openstack-operators(90e1030d-d9d0-42bf-b35b-a2f19ee0e6de): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.029954 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j"] Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.031603 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bfx24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-86d796d84d-vs899_openstack-operators(a44bf58b-6eb3-4bdc-ad1c-41aa7623666b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.038642 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp"] Nov 24 13:59:54 crc kubenswrapper[4970]: W1124 13:59:54.041524 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod289cc484_057a_4107_937d_6c532c21b52a.slice/crio-250ee4111f5951dfa4d78eb8b7b5ffc17415ea0b6792b39e76c04d5fab793a66 WatchSource:0}: Error finding container 250ee4111f5951dfa4d78eb8b7b5ffc17415ea0b6792b39e76c04d5fab793a66: Status 404 returned error can't find the container with id 250ee4111f5951dfa4d78eb8b7b5ffc17415ea0b6792b39e76c04d5fab793a66 Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.044346 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-slrrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7798859c74-k5b5j_openstack-operators(289cc484-057a-4107-937d-6c532c21b52a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.045496 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-86d796d84d-vs899"] Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.045900 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pl7qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-6fdc856c5d-g8kmd_openstack-operators(d94c9e13-b160-4f65-a97c-9bfdd4799eb0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.055443 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-76hbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5bdf4f7f7f-hdflg_openstack-operators(1af43cf6-5835-47eb-a315-c16bf5758a9e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.061438 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd"] Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.215203 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l"] Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.221475 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd"] Nov 24 13:59:54 crc kubenswrapper[4970]: W1124 13:59:54.223406 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9847f51b_ce73_493a_ac47_b93341cebf73.slice/crio-7f748c7af969a2aebfd3af0cfdc4e64abef38185ba92c1e77b480dbacc15076d WatchSource:0}: Error finding container 7f748c7af969a2aebfd3af0cfdc4e64abef38185ba92c1e77b480dbacc15076d: Status 404 returned error can't find the container with id 7f748c7af969a2aebfd3af0cfdc4e64abef38185ba92c1e77b480dbacc15076d Nov 24 13:59:54 crc kubenswrapper[4970]: W1124 13:59:54.225898 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65008609_c137_4dee_b406_159814902ff4.slice/crio-ff7ba379dcece5fe4c9b7a4690412424a8ab5a2a49da1284706a28254b805871 WatchSource:0}: Error finding container ff7ba379dcece5fe4c9b7a4690412424a8ab5a2a49da1284706a28254b805871: Status 404 returned error can't find the container with id ff7ba379dcece5fe4c9b7a4690412424a8ab5a2a49da1284706a28254b805871 Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.228643 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bd2wh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7cd4fb6f79-m9m2l_openstack-operators(9847f51b-ce73-493a-ac47-b93341cebf73): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.257120 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp"] Nov 24 13:59:54 crc kubenswrapper[4970]: W1124 13:59:54.266999 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd17a257a_f687_4d53_9b81_5fa430e24d70.slice/crio-0333052ed0e50bbb70f9ca4b97941922856b7f4abfbf8d582fc6a7ee5ca7fb7c WatchSource:0}: Error finding container 0333052ed0e50bbb70f9ca4b97941922856b7f4abfbf8d582fc6a7ee5ca7fb7c: Status 404 returned error can't find the container with id 0333052ed0e50bbb70f9ca4b97941922856b7f4abfbf8d582fc6a7ee5ca7fb7c Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.270876 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:7dbadf7b98f2f305f9f1382f55a084c8ca404f4263f76b28e56bd0dc437e2192,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:c8101c77a82eae4407e41e1fd766dfc6e1b7f9ed1679e3efb6f91ff97a1557b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:eb9743b21bbadca6f7cb9ac4fc46b5d58c51c674073c7e1121f4474a71304071,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:3d81f839b98c2e2a5bf0da79f2f9a92dff7d0a3c5a830b0e95c89dad8cf98a6a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:d19ac99249b47dd8ea16cd6aaa5756346aa8a2f119ee50819c15c5366efb417d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:8536169e5537fe6c330eba814248abdcf39cdd8f7e7336034d74e6fda9544050,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:4f1fa337760e82bfd67cdd142a97c121146dd7e621daac161940dd5e4ddb80dc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:3613b345d5baed98effd906f8b0242d863e14c97078ea473ef01fe1b0afc46f3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:d375d370be5ead0dac71109af644849e5795f535f9ad8eeacea261d77ae6f140,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:9f9f367ed4c85efb16c3a74a4bb707ff0db271d7bc5abc70a71e984b55f43003,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:b73ad22b4955b06d584bce81742556d8c0c7828c495494f8ea7c99391c61b70f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:aa1d3aaf6b394621ed4089a98e0a82b763f467e8b5c5db772f9fdf99fc86e333,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:d6661053141b6df421288a7c9968a155ab82e478c1d75ab41f2cebe2f0ca02d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:ce2d63258cb4e7d0d1c07234de6889c5434464190906798019311a1c7cf6387f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:0485ef9e5b4437f7cd2ba54034a87722ce4669ee86b3773c6b0c037ed8000e91,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api@sha256:962c004551d0503779364b767b9bf0cecdf78dbba8809b2ca8b073f58e1f4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor@sha256:0ebf4c465fb6cc7dad9e6cb2da0ff54874c9acbcb40d62234a629ec2c12cdd62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:ff0c553ceeb2e0f44b010e37dc6d0db8a251797b88e56468b7cf7f05253e4232,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:624f553f073af7493d34828b074adc9981cce403edd8e71482c7307008479fd9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:e3874936a518c8560339db8f840fc5461885819f6050b5de8d3ab9199bea5094,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:1cea25f1d2a45affc80c46fb9d427749d3f06b61590ac6070a2910e3ec8a4e5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:e36d5b9a65194f12f7b01c6422ba3ed52a687fd1695fbb21f4986c67d9f9317f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:8b21bec527d54cd766e277889df6bcccd2baeaa946274606b986c0c3b7ca689f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:45aceca77f8fcf61127f0da650bdfdf11ede9b0944c78b63fab819d03283f96b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:709ac58998927dd61786821ae1e63343fd97ccf5763aac5edb4583eea9401d22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:867d4ef7c21f75e6030a685b5762ab4d84b671316ed6b98d75200076e93342cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2b90da93550b99d2fcfa95bd819f3363aa68346a416f8dc7baac3e9c5f487761,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:6f86db36d668348be8c5b46dcda8b1fa23d34bfdc07164fbcbe7a6327fb4de24,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:8cde52cef8795d1c91983b100d86541c7718160ec260fe0f97b96add4c2c8ee8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:a9583cb3baf440d2358ef041373833afbeae60da8159dd031502379901141620,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:835ebed082fe1c45bd799d1d5357595ce63efeb05ca876f26b08443facb9c164,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:011d682241db724bc40736c9b54d2ea450ea7e6be095b1ff5fa28c8007466775,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:2025da90cff8f563deb08bee71efe16d4078edc2a767b2e225cca5c77f1aa2f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:26bd7b0bd6070856aefef6fe754c547d55c056396ea30d879d34c2d49b5a1d29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:ff46cd5e0e13d105c4629e78c2734a50835f06b6a1e31da9e0462981d10c4be3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:5b4fd0c2b76fa5539f74687b11c5882d77bd31352452322b37ff51fa18f12a61,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:5f6045841aff0fde6f684a34cdf49f8dc7b2c3bcbdeab201f1058971e0c5f79e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:448f4e1b740c30936e340bd6e8534d78c83357bf373a4223950aa64d3484f007,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:b68e3615af8a0eb0ef6bf9ceeef59540a6f4a9a85f6078a3620be115c73a7db8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:7eae01cf60383e523c9cd94d158a9162120a7370829a1dad20fdea6b0fd660bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:28cc10501788081eb61b5a1af35546191a92741f4f109df54c74e2b19439d0f9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:9a616e37acfd120612f78043237a8541266ba34883833c9beb43f3da313661ad,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:6b1be6cd94a0942259bca5d5d2c30cc7de4a33276b61f8ae3940226772106256,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:02d2c22d15401574941fbe057095442dee0d6f7a0a9341de35d25e6a12a3fe4b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:fc3b3a36b74fd653946723c54b208072d52200635850b531e9d595a7aaea5a01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:7850ccbff320bf9a1c9c769c1c70777eb97117dd8cd5ae4435be9b4622cf807a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:397dac7e39cf40d14a986e6ec4a60fb698ca35c197d0db315b1318514cc6d1d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:10452e2144368e2f128c8fb8ef9e54880b06ef1d71d9f084a0217dcb099c51ce,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:1c95142a36276686e720f86423ee171dc9adcc1e89879f627545b7c906ccd9bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:e331a8fde6638e5ba154c4f0b38772a9a424f60656f2777245975fb1fa02f07d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:b6e1e8a249d36ef36c6ac4170af1e043dda1ccc0f9672832d3ff151bf3533076,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:cd3cf7a34053e850b4d4f9f4ea4c74953a54a42fd18e47d7c01d44a88923e925,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:aee28476344fc0cc148fbe97daf9b1bfcedc22001550bba4bdc4e84be7b6989d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:cfa0b92c976603ee2a937d34013a238fcd8aa75f998e50642e33489f14124633,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:73c2f2d6eecf88acf4e45b133c8373d9bb006b530e0aff0b28f3b7420620a874,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:927b405cc04abe5ff716186e8d35e2dc5fad1c8430194659ee6617d74e4e055d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:6154d7cebd7c339afa5b86330262156171743aa5b79c2b78f9a2f378005ed8fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:e2db2f4af8d3d0be7868c6efef0189f3a2c74a8f96ae10e3f991cdf83feaef29,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:c773629df257726a6d3cacc24a6e4df0babcd7d37df04e6d14676a8da028b9c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:776211111e2e6493706dbc49a3ba44f31d1b947919313ed3a0f35810e304ec52,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:0a98e8f5c83522ca6c8e40c5e9561f6628d2d5e69f0e8a64279c541c989d3d8b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:7cccf24ad0a152f90ca39893064f48a1656950ee8142685a5d482c71f0bdc9f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:af46761060c7987e1dee5f14c06d85b46f12ad8e09c83d4246ab4e3a65dfda3e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:05450b48f6b5352b2686a26e933e8727748edae2ae9652d9164b7d7a1817c55a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:fc9c99eeef91523482bd8f92661b393287e1f2a24ad2ba9e33191f8de9af74cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:3e4ecc02b4b5e0860482a93599ba9ca598c5ce26c093c46e701f96fe51acb208,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:2346037e064861c7892690d2e8b3e1eea1a26ce3c3a11fda0b41301965bc828c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:c26c3ff9cabe3593ceb10006e782bf9391ac14785768ce9eec4f938c2d3cf228,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:273fe8c27d08d0f62773a02f8cef6a761a7768116ee1a4be611f93bbf63f2b75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:daa45220bb1c47922d0917aa8fe423bb82b03a01429f1c9e37635e701e352d71,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:a80a074e227d3238bb6f285788a9e886ae7a5909ccbc5c19c93c369bdfe5b3b8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:58ac66ca1be01fe0157977bd79a26cde4d0de153edfaf4162367c924826b2ef4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:99a63770d80cc7c3afa1118b400972fb0e6bff5284a2eae781b12582ad79c29c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:9ee4d84529394afcd860f1a1186484560f02f08c15c37cac42a22473b7116d5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:ea15fadda7b0439ec637edfaf6ea5dbf3e35fb3be012c7c5a31e722c90becb11,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zd5g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp_openstack-operators(d17a257a-f687-4d53-9b81-5fa430e24d70): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.342800 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx"] Nov 24 13:59:54 crc kubenswrapper[4970]: W1124 13:59:54.353754 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ed5751d_33ce_4775_a0c7_cb951a6fbdcb.slice/crio-6453cda074bd89584620c0469406221711e8f5ca61ebfc809ad008d05623f637 WatchSource:0}: Error finding container 6453cda074bd89584620c0469406221711e8f5ca61ebfc809ad008d05623f637: Status 404 returned error can't find the container with id 6453cda074bd89584620c0469406221711e8f5ca61ebfc809ad008d05623f637 Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.432405 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" podUID="90e1030d-d9d0-42bf-b35b-a2f19ee0e6de" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.457440 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" podUID="1af43cf6-5835-47eb-a315-c16bf5758a9e" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.465329 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" podUID="289cc484-057a-4107-937d-6c532c21b52a" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.477452 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" podUID="9847f51b-ce73-493a-ac47-b93341cebf73" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.478011 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" podUID="a44bf58b-6eb3-4bdc-ad1c-41aa7623666b" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.482843 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" podUID="d17a257a-f687-4d53-9b81-5fa430e24d70" Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.508344 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" podUID="d94c9e13-b160-4f65-a97c-9bfdd4799eb0" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.591337 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" event={"ID":"b070979e-cf9f-4793-a50a-f0e67a48a58f","Type":"ContainerStarted","Data":"df4f23c18efbff1fb6520cffe03cce2eef0f1fa06af210183d1e88122941d6ba"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.601017 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" event={"ID":"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b","Type":"ContainerStarted","Data":"614155055b062c0379d499e8a2bfd270bc5c5bd235df4331c337ac13a2b1639f"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.601056 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" event={"ID":"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b","Type":"ContainerStarted","Data":"f9581436cc3e60098758e02b0349c015febc67d3334d6017bcc0d067f5777d0c"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.606786 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" event={"ID":"73f36b2f-3ac8-4d5c-859b-1e376b9a5194","Type":"ContainerStarted","Data":"63d4b57a52497888e13651956e3d77baa5bbc21f7fb82962f8a74321e27be94d"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.609409 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" podUID="a44bf58b-6eb3-4bdc-ad1c-41aa7623666b" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.612858 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" event={"ID":"289cc484-057a-4107-937d-6c532c21b52a","Type":"ContainerStarted","Data":"9545b6f1b4d839c3ed4c81c877003a1be25efa0e3974126a6c5a684f2106e7ed"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.612904 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" event={"ID":"289cc484-057a-4107-937d-6c532c21b52a","Type":"ContainerStarted","Data":"250ee4111f5951dfa4d78eb8b7b5ffc17415ea0b6792b39e76c04d5fab793a66"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.614082 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" podUID="289cc484-057a-4107-937d-6c532c21b52a" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.615126 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" event={"ID":"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7","Type":"ContainerStarted","Data":"f9bb7b129d760f0f5aa97b01fb22f0af7889a3f443bbb400fd2de165c79607ee"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.620856 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" event={"ID":"a5f4428d-ecd0-4c71-b9da-de188497a191","Type":"ContainerStarted","Data":"998c3d1ee574e47f94847ca4c8fca3eb109e6be5097281add8e8ce8cfae35519"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.626785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" event={"ID":"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb","Type":"ContainerStarted","Data":"b7b267ec3d84847ad132dcb4f76cc816b6c7cc018ef7c27e898cf266c3642fe5"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.626829 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" event={"ID":"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb","Type":"ContainerStarted","Data":"6453cda074bd89584620c0469406221711e8f5ca61ebfc809ad008d05623f637"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.630325 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" event={"ID":"20a43498-16e1-46d8-9487-b87ff9cad8f8","Type":"ContainerStarted","Data":"130bebb046ba964577376fa3c7d9a2364b0c4576ffe31b904a4bedd76ae89e04"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.639886 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" event={"ID":"d94c9e13-b160-4f65-a97c-9bfdd4799eb0","Type":"ContainerStarted","Data":"664a7ca375cd68e3e2120aaf65b2cb42b23035e7640a3442d88879c8c2168c45"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.639928 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" event={"ID":"d94c9e13-b160-4f65-a97c-9bfdd4799eb0","Type":"ContainerStarted","Data":"c153211349244970c4f3bd5bea6482887f7586ef1cf552e226695f665773c637"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.647066 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" podUID="d94c9e13-b160-4f65-a97c-9bfdd4799eb0" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.648114 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" event={"ID":"65008609-c137-4dee-b406-159814902ff4","Type":"ContainerStarted","Data":"ff7ba379dcece5fe4c9b7a4690412424a8ab5a2a49da1284706a28254b805871"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.652867 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" event={"ID":"9847f51b-ce73-493a-ac47-b93341cebf73","Type":"ContainerStarted","Data":"22c265301abad62e9f10da9bab6d847b489cf994b1d4d414d4335fc022549244"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.652957 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" event={"ID":"9847f51b-ce73-493a-ac47-b93341cebf73","Type":"ContainerStarted","Data":"7f748c7af969a2aebfd3af0cfdc4e64abef38185ba92c1e77b480dbacc15076d"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.654047 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" podUID="9847f51b-ce73-493a-ac47-b93341cebf73" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.656154 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" event={"ID":"d17a257a-f687-4d53-9b81-5fa430e24d70","Type":"ContainerStarted","Data":"897328d33a20ba41a3a28f2481aba7ce5d0c7e93faafc4cf215af6797b84a896"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.656207 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" event={"ID":"d17a257a-f687-4d53-9b81-5fa430e24d70","Type":"ContainerStarted","Data":"0333052ed0e50bbb70f9ca4b97941922856b7f4abfbf8d582fc6a7ee5ca7fb7c"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.660577 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" podUID="d17a257a-f687-4d53-9b81-5fa430e24d70" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.660785 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" event={"ID":"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de","Type":"ContainerStarted","Data":"e769ff3c5e6103be7781aadf5512ae90b25e87a274c2d745b01b2906c5ffe6e8"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.660839 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" event={"ID":"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de","Type":"ContainerStarted","Data":"a959563e67c39774ad9737a7e69f80c11fbc06e0beb77a0732d6e6b3649599d9"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.663767 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" event={"ID":"028f94a9-1db9-46e8-93f1-106d959ebe9b","Type":"ContainerStarted","Data":"77195e189c97bfa9e10815b26ef81daf85bbe2007d67a2697421386edb4608db"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.667025 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" podUID="90e1030d-d9d0-42bf-b35b-a2f19ee0e6de" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.668847 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" event={"ID":"f01cc05b-92d6-4144-ad87-5fdec63cc95c","Type":"ContainerStarted","Data":"237624fed825cd1fe12c247b8cfb5222d309ad2ac06ffb9283c02a4bc74ae248"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.673716 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" event={"ID":"1af43cf6-5835-47eb-a315-c16bf5758a9e","Type":"ContainerStarted","Data":"62c6605792709091aff4e926c7b7cbd24ffcf3a8e0fdbdcf0a56cc3a88205927"} Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.674422 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" event={"ID":"1af43cf6-5835-47eb-a315-c16bf5758a9e","Type":"ContainerStarted","Data":"85f5b3bc9860bedb0d5610a02b658152f7233a977a4132e98949a7abaed8614d"} Nov 24 13:59:54 crc kubenswrapper[4970]: E1124 13:59:54.675935 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" podUID="1af43cf6-5835-47eb-a315-c16bf5758a9e" Nov 24 13:59:54 crc kubenswrapper[4970]: I1124 13:59:54.677191 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" event={"ID":"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20","Type":"ContainerStarted","Data":"1c177da65894da14f52c133c18538115d6fc1dce51194a79c34ff836345ba500"} Nov 24 13:59:55 crc kubenswrapper[4970]: I1124 13:59:55.697351 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" event={"ID":"6ed5751d-33ce-4775-a0c7-cb951a6fbdcb","Type":"ContainerStarted","Data":"7ad7d43d78ab53cea2ac2d75df8a41d7f0b9d23c435d92091b775c296e2e5e22"} Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700676 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" podUID="d17a257a-f687-4d53-9b81-5fa430e24d70" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700676 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\"" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" podUID="a44bf58b-6eb3-4bdc-ad1c-41aa7623666b" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700724 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" podUID="d94c9e13-b160-4f65-a97c-9bfdd4799eb0" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700757 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" podUID="90e1030d-d9d0-42bf-b35b-a2f19ee0e6de" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700786 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" podUID="289cc484-057a-4107-937d-6c532c21b52a" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.700815 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" podUID="9847f51b-ce73-493a-ac47-b93341cebf73" Nov 24 13:59:55 crc kubenswrapper[4970]: E1124 13:59:55.702328 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" podUID="1af43cf6-5835-47eb-a315-c16bf5758a9e" Nov 24 13:59:55 crc kubenswrapper[4970]: I1124 13:59:55.753174 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" podStartSLOduration=3.753139897 podStartE2EDuration="3.753139897s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 13:59:55.748064675 +0000 UTC m=+811.035821958" watchObservedRunningTime="2025-11-24 13:59:55.753139897 +0000 UTC m=+811.040897190" Nov 24 13:59:56 crc kubenswrapper[4970]: I1124 13:59:56.705053 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.137791 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7"] Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.139310 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.144112 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.144637 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.149764 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7"] Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.247682 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhr5p\" (UniqueName: \"kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.247735 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.247785 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.349355 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.349485 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhr5p\" (UniqueName: \"kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.349524 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.351480 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.355770 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.365001 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhr5p\" (UniqueName: \"kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p\") pod \"collect-profiles-29399880-zmbj7\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:00 crc kubenswrapper[4970]: I1124 14:00:00.472995 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:03 crc kubenswrapper[4970]: I1124 14:00:03.883156 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6cb9dc54f8-bcbxx" Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.109094 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7"] Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.905813 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" event={"ID":"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7","Type":"ContainerStarted","Data":"0ca3d5a0a308e4807ecbe801ee8f5f11c9b7794b8f84edae9ad416146491702e"} Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.957028 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" event={"ID":"a5f4428d-ecd0-4c71-b9da-de188497a191","Type":"ContainerStarted","Data":"f58163b1aa4abd53d2eef572cb9ba0af479e3f53c99fdb68632e204476ea77b5"} Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.967083 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" event={"ID":"799af0bf-24f0-4546-960b-2e8e72083fb8","Type":"ContainerStarted","Data":"01fcdc0c74f9555d42d2bdef7fba2f0cc206b759b69b92155ecec53648883906"} Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.982860 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" event={"ID":"73f36b2f-3ac8-4d5c-859b-1e376b9a5194","Type":"ContainerStarted","Data":"b2594001df15daf7bae78c127a27665b104ed0450cca86e256a46c02c2c51c74"} Nov 24 14:00:04 crc kubenswrapper[4970]: I1124 14:00:04.992019 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" event={"ID":"65008609-c137-4dee-b406-159814902ff4","Type":"ContainerStarted","Data":"9876e89240070022ffd1fc91f539e53c01892762c46bafd89898216604060fb9"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.009919 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" event={"ID":"9ca1ba14-aafe-429b-be3d-87dadeb73b54","Type":"ContainerStarted","Data":"0d3135823d98d1f29580f3652b1ff625ca119616817384beaef7e70989601aad"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.017786 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" event={"ID":"028f94a9-1db9-46e8-93f1-106d959ebe9b","Type":"ContainerStarted","Data":"b3b972fbe964f20ce15ad1c3d72316d6d1bbdd3cb283f0914c082122d4ce900d"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.017827 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" event={"ID":"028f94a9-1db9-46e8-93f1-106d959ebe9b","Type":"ContainerStarted","Data":"87ef50d8142942c2d978de876b04b536a93f4829ae62eb337997c01058f59d6c"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.018662 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.019617 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-488cd" podStartSLOduration=3.528731725 podStartE2EDuration="13.019597539s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.228149514 +0000 UTC m=+809.515906807" lastFinishedPulling="2025-11-24 14:00:03.719015328 +0000 UTC m=+819.006772621" observedRunningTime="2025-11-24 14:00:05.015871311 +0000 UTC m=+820.303628604" watchObservedRunningTime="2025-11-24 14:00:05.019597539 +0000 UTC m=+820.307354832" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.021337 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" event={"ID":"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd","Type":"ContainerStarted","Data":"7194698f300fe27cba801649b1b21c751fdd9e8b99a1382dbab7eae85f1383d8"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.027038 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" event={"ID":"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0","Type":"ContainerStarted","Data":"c7dc8db4da62ad3609f6827eb91366da49a9f86569f2226281494b6e8e7c5692"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.027087 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" event={"ID":"d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0","Type":"ContainerStarted","Data":"1fc3fadf4ace4d90c2370e387ffb33f181b41115f8b3ddf00cd9b9c5e969f0f5"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.027368 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.038161 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" event={"ID":"8874383f-46c3-43b7-b1d1-56c9d14a0004","Type":"ContainerStarted","Data":"ed3d28b6f6cac333da0c600ce74f9705678058baba9d596878ef1490526f39d3"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.038194 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" event={"ID":"8874383f-46c3-43b7-b1d1-56c9d14a0004","Type":"ContainerStarted","Data":"d3b9809d79e66c8d0263380a5011e8a120280ae56aaef71a0924578a9b7d14de"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.038585 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.042679 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" event={"ID":"20a43498-16e1-46d8-9487-b87ff9cad8f8","Type":"ContainerStarted","Data":"04466cbccb61a3223c8d58a857944529dec3cec7da2b9f93c00ba130d361617b"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.049171 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" podStartSLOduration=3.469023991 podStartE2EDuration="13.049150788s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.028808755 +0000 UTC m=+809.316566048" lastFinishedPulling="2025-11-24 14:00:03.608935542 +0000 UTC m=+818.896692845" observedRunningTime="2025-11-24 14:00:05.040300787 +0000 UTC m=+820.328058080" watchObservedRunningTime="2025-11-24 14:00:05.049150788 +0000 UTC m=+820.336908081" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.055726 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" event={"ID":"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20","Type":"ContainerStarted","Data":"bc953089d124b1542d306e3e7f34ae3efc8bf279cedc1e2aad78e02f491a5784"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.072607 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" event={"ID":"b070979e-cf9f-4793-a50a-f0e67a48a58f","Type":"ContainerStarted","Data":"2dc35359addd62ef94b7a751c825bea33f8a68b88e5fd93f796080132e9dbbce"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.075172 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" event={"ID":"cc36780e-1f96-45c0-a50b-fd9befc47823","Type":"ContainerStarted","Data":"d331b7226f2105249a2f1f2e8ae18e2734b70a803266573fd40eb381fdbcfb2e"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.079363 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" event={"ID":"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5","Type":"ContainerStarted","Data":"13b3251a3ff12f7a00f63521561d98442204cbadfd88a077fca074f817e61f09"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.081352 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" podStartSLOduration=4.000089083 podStartE2EDuration="14.081332386s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.540712147 +0000 UTC m=+808.828469430" lastFinishedPulling="2025-11-24 14:00:03.62195543 +0000 UTC m=+818.909712733" observedRunningTime="2025-11-24 14:00:05.077836744 +0000 UTC m=+820.365594037" watchObservedRunningTime="2025-11-24 14:00:05.081332386 +0000 UTC m=+820.369089679" Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.082641 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" event={"ID":"f01cc05b-92d6-4144-ad87-5fdec63cc95c","Type":"ContainerStarted","Data":"9f548833d853653eb76bfe2bec6a927c46ae6c5209424b01df8b021bc0abd2a4"} Nov 24 14:00:05 crc kubenswrapper[4970]: I1124 14:00:05.095073 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" podStartSLOduration=3.749265682 podStartE2EDuration="14.095049733s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.28002441 +0000 UTC m=+808.567781703" lastFinishedPulling="2025-11-24 14:00:03.625808461 +0000 UTC m=+818.913565754" observedRunningTime="2025-11-24 14:00:05.091185812 +0000 UTC m=+820.378943105" watchObservedRunningTime="2025-11-24 14:00:05.095049733 +0000 UTC m=+820.382807036" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.116876 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" event={"ID":"20a43498-16e1-46d8-9487-b87ff9cad8f8","Type":"ContainerStarted","Data":"6eb31ecfdb6c25b1472d4ccab4e65218f9c216e708e243792242f66f0ac0f211"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.118274 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.121334 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" event={"ID":"0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20","Type":"ContainerStarted","Data":"12642bd326f20666d898e30e9a5def0f77e5f524cfa9b93f6f127a85052c2acd"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.121473 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.123964 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" event={"ID":"b070979e-cf9f-4793-a50a-f0e67a48a58f","Type":"ContainerStarted","Data":"74a69cf7ba87945e08abd1c797c695ffb0206a4488235d629ca7596c28d8f7b9"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.124088 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.126226 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" event={"ID":"4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd","Type":"ContainerStarted","Data":"313aabdfd072fc5c68d7be5b510e9fe8f69aadc556596a19ae0ad14f5b7a9f67"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.126587 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.129721 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" event={"ID":"73f36b2f-3ac8-4d5c-859b-1e376b9a5194","Type":"ContainerStarted","Data":"e857c4f7627c8450638f3e2d28ba5a18a2f3893117775468739c1ffad2d776f9"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.129939 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.132067 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" event={"ID":"93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5","Type":"ContainerStarted","Data":"a167a00e8a8173bd9c35ebbf4df828973f29b9334fde7d38ea10c2a649a9ed10"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.132146 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.137836 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" podStartSLOduration=4.534070658 podStartE2EDuration="14.137817401s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.019767159 +0000 UTC m=+809.307524442" lastFinishedPulling="2025-11-24 14:00:03.623513892 +0000 UTC m=+818.911271185" observedRunningTime="2025-11-24 14:00:06.132278947 +0000 UTC m=+821.420036230" watchObservedRunningTime="2025-11-24 14:00:06.137817401 +0000 UTC m=+821.425574694" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.140219 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" event={"ID":"9ca1ba14-aafe-429b-be3d-87dadeb73b54","Type":"ContainerStarted","Data":"1cabd59a419531282dbcb0619f8d9b67ed92dab788fd1029a61714e772fe5419"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.140277 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.144948 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" event={"ID":"5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7","Type":"ContainerStarted","Data":"c44af2b2cb021d48de365516dc63bfcd45ce051d11fc416255a7f18002a61733"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.145648 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.151351 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" podStartSLOduration=4.550942728 podStartE2EDuration="14.151330503s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.023487996 +0000 UTC m=+809.311245289" lastFinishedPulling="2025-11-24 14:00:03.623875761 +0000 UTC m=+818.911633064" observedRunningTime="2025-11-24 14:00:06.147021801 +0000 UTC m=+821.434779134" watchObservedRunningTime="2025-11-24 14:00:06.151330503 +0000 UTC m=+821.439087806" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.154796 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" event={"ID":"a5f4428d-ecd0-4c71-b9da-de188497a191","Type":"ContainerStarted","Data":"01ad6a77638e13fcd5050f5ca0bd9de201dde3bad3c79d751c67740dd05e09b6"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.154955 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.156733 4970 generic.go:334] "Generic (PLEG): container finished" podID="cc36780e-1f96-45c0-a50b-fd9befc47823" containerID="c83804b15721950704e8f7ab50433d08eedd48449afd5755fc56244e84fccf48" exitCode=0 Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.156791 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" event={"ID":"cc36780e-1f96-45c0-a50b-fd9befc47823","Type":"ContainerDied","Data":"c83804b15721950704e8f7ab50433d08eedd48449afd5755fc56244e84fccf48"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.158710 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" event={"ID":"799af0bf-24f0-4546-960b-2e8e72083fb8","Type":"ContainerStarted","Data":"23015ac49bfb383b8a661e657ccb9b16a17bdf9dc0a503b62c21ccecf9bdce4b"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.158851 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.160641 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" event={"ID":"f01cc05b-92d6-4144-ad87-5fdec63cc95c","Type":"ContainerStarted","Data":"ed90f6c702179bcad8e2b0b09095dd77492aa1e0a4ea96d57da03793a6fcad6e"} Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.162394 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.177819 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" podStartSLOduration=4.408288599 podStartE2EDuration="15.177790022s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:52.852868909 +0000 UTC m=+808.140626202" lastFinishedPulling="2025-11-24 14:00:03.622370312 +0000 UTC m=+818.910127625" observedRunningTime="2025-11-24 14:00:06.169412094 +0000 UTC m=+821.457169387" watchObservedRunningTime="2025-11-24 14:00:06.177790022 +0000 UTC m=+821.465547355" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.185747 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" podStartSLOduration=4.919752615 podStartE2EDuration="15.185730399s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.359893939 +0000 UTC m=+808.647651222" lastFinishedPulling="2025-11-24 14:00:03.625871703 +0000 UTC m=+818.913629006" observedRunningTime="2025-11-24 14:00:06.185224435 +0000 UTC m=+821.472981728" watchObservedRunningTime="2025-11-24 14:00:06.185730399 +0000 UTC m=+821.473487692" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.209097 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" podStartSLOduration=4.162434712 podStartE2EDuration="14.209078206s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.574105276 +0000 UTC m=+808.861862559" lastFinishedPulling="2025-11-24 14:00:03.62074875 +0000 UTC m=+818.908506053" observedRunningTime="2025-11-24 14:00:06.205114763 +0000 UTC m=+821.492872056" watchObservedRunningTime="2025-11-24 14:00:06.209078206 +0000 UTC m=+821.496835499" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.229488 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" podStartSLOduration=4.626031133 podStartE2EDuration="14.229470268s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.020059457 +0000 UTC m=+809.307816750" lastFinishedPulling="2025-11-24 14:00:03.623498572 +0000 UTC m=+818.911255885" observedRunningTime="2025-11-24 14:00:06.224077617 +0000 UTC m=+821.511834910" watchObservedRunningTime="2025-11-24 14:00:06.229470268 +0000 UTC m=+821.517227561" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.243073 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" podStartSLOduration=5.19266999 podStartE2EDuration="15.24305233s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.573419729 +0000 UTC m=+808.861177022" lastFinishedPulling="2025-11-24 14:00:03.623802069 +0000 UTC m=+818.911559362" observedRunningTime="2025-11-24 14:00:06.241426668 +0000 UTC m=+821.529183961" watchObservedRunningTime="2025-11-24 14:00:06.24305233 +0000 UTC m=+821.530809623" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.257260 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" podStartSLOduration=4.907124237 podStartE2EDuration="15.257242371s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.273642244 +0000 UTC m=+808.561399537" lastFinishedPulling="2025-11-24 14:00:03.623760378 +0000 UTC m=+818.911517671" observedRunningTime="2025-11-24 14:00:06.254657813 +0000 UTC m=+821.542415116" watchObservedRunningTime="2025-11-24 14:00:06.257242371 +0000 UTC m=+821.544999664" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.274035 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" podStartSLOduration=5.024758509 podStartE2EDuration="15.274016297s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.374462629 +0000 UTC m=+808.662219922" lastFinishedPulling="2025-11-24 14:00:03.623720407 +0000 UTC m=+818.911477710" observedRunningTime="2025-11-24 14:00:06.273811821 +0000 UTC m=+821.561569114" watchObservedRunningTime="2025-11-24 14:00:06.274016297 +0000 UTC m=+821.561773590" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.296449 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" podStartSLOduration=4.670409759 podStartE2EDuration="14.296422231s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.996252537 +0000 UTC m=+809.284009830" lastFinishedPulling="2025-11-24 14:00:03.622264999 +0000 UTC m=+818.910022302" observedRunningTime="2025-11-24 14:00:06.29182251 +0000 UTC m=+821.579579823" watchObservedRunningTime="2025-11-24 14:00:06.296422231 +0000 UTC m=+821.584179534" Nov 24 14:00:06 crc kubenswrapper[4970]: I1124 14:00:06.311698 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" podStartSLOduration=5.285764144 podStartE2EDuration="15.311683027s" podCreationTimestamp="2025-11-24 13:59:51 +0000 UTC" firstStartedPulling="2025-11-24 13:59:53.597393793 +0000 UTC m=+808.885151086" lastFinishedPulling="2025-11-24 14:00:03.623312666 +0000 UTC m=+818.911069969" observedRunningTime="2025-11-24 14:00:06.308195767 +0000 UTC m=+821.595953060" watchObservedRunningTime="2025-11-24 14:00:06.311683027 +0000 UTC m=+821.599440320" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.487477 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.539517 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:07 crc kubenswrapper[4970]: E1124 14:00:07.539829 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc36780e-1f96-45c0-a50b-fd9befc47823" containerName="collect-profiles" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.539845 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc36780e-1f96-45c0-a50b-fd9befc47823" containerName="collect-profiles" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.539991 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc36780e-1f96-45c0-a50b-fd9befc47823" containerName="collect-profiles" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.541561 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.550098 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.578198 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhr5p\" (UniqueName: \"kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p\") pod \"cc36780e-1f96-45c0-a50b-fd9befc47823\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.578338 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.578388 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vh4g\" (UniqueName: \"kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.578464 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.589446 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p" (OuterVolumeSpecName: "kube-api-access-dhr5p") pod "cc36780e-1f96-45c0-a50b-fd9befc47823" (UID: "cc36780e-1f96-45c0-a50b-fd9befc47823"). InnerVolumeSpecName "kube-api-access-dhr5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679505 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume\") pod \"cc36780e-1f96-45c0-a50b-fd9befc47823\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679669 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume\") pod \"cc36780e-1f96-45c0-a50b-fd9befc47823\" (UID: \"cc36780e-1f96-45c0-a50b-fd9befc47823\") " Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679813 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vh4g\" (UniqueName: \"kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679873 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679922 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.679957 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhr5p\" (UniqueName: \"kubernetes.io/projected/cc36780e-1f96-45c0-a50b-fd9befc47823-kube-api-access-dhr5p\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.681266 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.681336 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.681504 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume" (OuterVolumeSpecName: "config-volume") pod "cc36780e-1f96-45c0-a50b-fd9befc47823" (UID: "cc36780e-1f96-45c0-a50b-fd9befc47823"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.684645 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cc36780e-1f96-45c0-a50b-fd9befc47823" (UID: "cc36780e-1f96-45c0-a50b-fd9befc47823"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.701463 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vh4g\" (UniqueName: \"kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g\") pod \"redhat-marketplace-xwjpn\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.782967 4970 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cc36780e-1f96-45c0-a50b-fd9befc47823-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.783215 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cc36780e-1f96-45c0-a50b-fd9befc47823-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:07 crc kubenswrapper[4970]: I1124 14:00:07.863893 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:08 crc kubenswrapper[4970]: I1124 14:00:08.178098 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" event={"ID":"cc36780e-1f96-45c0-a50b-fd9befc47823","Type":"ContainerDied","Data":"d331b7226f2105249a2f1f2e8ae18e2734b70a803266573fd40eb381fdbcfb2e"} Nov 24 14:00:08 crc kubenswrapper[4970]: I1124 14:00:08.178139 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d331b7226f2105249a2f1f2e8ae18e2734b70a803266573fd40eb381fdbcfb2e" Nov 24 14:00:08 crc kubenswrapper[4970]: I1124 14:00:08.178161 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.202079 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7768f8c84f-kdx5m" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.285150 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-56dfb6b67f-ff65l" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.313899 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6d8fd67bf7-r452c" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.322631 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-bf4c6585d-v67fx" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.360345 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-8667fbf6f6-qclm2" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.394665 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d86b44686-crqqj" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.416882 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-769d9c7585-9lhqx" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.504401 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5c75d7c94b-sx2v5" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.572051 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7bb88cb858-h9j24" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.602790 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7879fb76fd-9mnkp" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.610350 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6f8c5b86cb-wfsc5" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.719482 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-66b7d6f598-zkcwk" Nov 24 14:00:12 crc kubenswrapper[4970]: I1124 14:00:12.884191 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-799cb6ffd6-5sp97" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.113259 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-8464cf66df-wfmzk" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.204441 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:13 crc kubenswrapper[4970]: W1124 14:00:13.210105 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1075b33_2cfe_4606_aa6d_53337a31bd9d.slice/crio-d9b48c3f292cb1b184192cede165de1a72cc9632c0abb9d043d4dc6dc6e8f962 WatchSource:0}: Error finding container d9b48c3f292cb1b184192cede165de1a72cc9632c0abb9d043d4dc6dc6e8f962: Status 404 returned error can't find the container with id d9b48c3f292cb1b184192cede165de1a72cc9632c0abb9d043d4dc6dc6e8f962 Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.218374 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" event={"ID":"d94c9e13-b160-4f65-a97c-9bfdd4799eb0","Type":"ContainerStarted","Data":"4bbe96fd6d7918458e4bc98c0ebcd24ef8b256ed50ba1a27f21d5446a0f9abc7"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.218592 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.222352 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" event={"ID":"9847f51b-ce73-493a-ac47-b93341cebf73","Type":"ContainerStarted","Data":"454130155fe288cb6ce02739c7ea5533dbead720c5565478a87f8eac57d5a232"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.222542 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.230889 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" event={"ID":"d17a257a-f687-4d53-9b81-5fa430e24d70","Type":"ContainerStarted","Data":"0cd55ceebb4f5d091b69bff68abc93b359dfad1dfc2728990cc49841b28205c8"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.231769 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.234754 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" event={"ID":"90e1030d-d9d0-42bf-b35b-a2f19ee0e6de","Type":"ContainerStarted","Data":"163529a3898529ed16d1c605b1cafc1091f53fc69f3bf9482349980ecee4c6c7"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.235404 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.237273 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" podStartSLOduration=2.49298867 podStartE2EDuration="21.237257026s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.045806497 +0000 UTC m=+809.333563790" lastFinishedPulling="2025-11-24 14:00:12.790074843 +0000 UTC m=+828.077832146" observedRunningTime="2025-11-24 14:00:13.234182086 +0000 UTC m=+828.521939379" watchObservedRunningTime="2025-11-24 14:00:13.237257026 +0000 UTC m=+828.525014319" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.239079 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" event={"ID":"1af43cf6-5835-47eb-a315-c16bf5758a9e","Type":"ContainerStarted","Data":"f705162156accb31c80cfb15a074a5191e5f08c8647091456e9d904d801acfa3"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.239233 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.240934 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" event={"ID":"289cc484-057a-4107-937d-6c532c21b52a","Type":"ContainerStarted","Data":"e422648a4947c5b096e16df7f13b7b1cb98908899e9948926279e530fc73c6e3"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.241094 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.242604 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" event={"ID":"a44bf58b-6eb3-4bdc-ad1c-41aa7623666b","Type":"ContainerStarted","Data":"cb959a0b6c6ca49e7f1a5a05a94c40c4e9b231b17f3da2d35cb7bd9ae1205899"} Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.242783 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.280927 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" podStartSLOduration=2.765784391 podStartE2EDuration="21.280909292s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.270169118 +0000 UTC m=+809.557926411" lastFinishedPulling="2025-11-24 14:00:12.785294009 +0000 UTC m=+828.073051312" observedRunningTime="2025-11-24 14:00:13.279769413 +0000 UTC m=+828.567526696" watchObservedRunningTime="2025-11-24 14:00:13.280909292 +0000 UTC m=+828.568666585" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.294981 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" podStartSLOduration=2.658015566 podStartE2EDuration="21.294964928s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.228501324 +0000 UTC m=+809.516258617" lastFinishedPulling="2025-11-24 14:00:12.865450686 +0000 UTC m=+828.153207979" observedRunningTime="2025-11-24 14:00:13.292772031 +0000 UTC m=+828.580529324" watchObservedRunningTime="2025-11-24 14:00:13.294964928 +0000 UTC m=+828.582722221" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.331024 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" podStartSLOduration=3.334600501 podStartE2EDuration="21.331010067s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.029068291 +0000 UTC m=+809.316825584" lastFinishedPulling="2025-11-24 14:00:12.025477857 +0000 UTC m=+827.313235150" observedRunningTime="2025-11-24 14:00:13.327268539 +0000 UTC m=+828.615025832" watchObservedRunningTime="2025-11-24 14:00:13.331010067 +0000 UTC m=+828.618767360" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.349792 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" podStartSLOduration=2.539956243 podStartE2EDuration="21.349767605s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.055316685 +0000 UTC m=+809.343073978" lastFinishedPulling="2025-11-24 14:00:12.865128047 +0000 UTC m=+828.152885340" observedRunningTime="2025-11-24 14:00:13.343555324 +0000 UTC m=+828.631312617" watchObservedRunningTime="2025-11-24 14:00:13.349767605 +0000 UTC m=+828.637524898" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.380025 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" podStartSLOduration=3.849494576 podStartE2EDuration="21.380006743s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.044123683 +0000 UTC m=+809.331880976" lastFinishedPulling="2025-11-24 14:00:11.57463585 +0000 UTC m=+826.862393143" observedRunningTime="2025-11-24 14:00:13.377109097 +0000 UTC m=+828.664866390" watchObservedRunningTime="2025-11-24 14:00:13.380006743 +0000 UTC m=+828.667764036" Nov 24 14:00:13 crc kubenswrapper[4970]: I1124 14:00:13.403495 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" podStartSLOduration=2.570486388 podStartE2EDuration="21.403474844s" podCreationTimestamp="2025-11-24 13:59:52 +0000 UTC" firstStartedPulling="2025-11-24 13:59:54.031452484 +0000 UTC m=+809.319209777" lastFinishedPulling="2025-11-24 14:00:12.86444094 +0000 UTC m=+828.152198233" observedRunningTime="2025-11-24 14:00:13.399951422 +0000 UTC m=+828.687708715" watchObservedRunningTime="2025-11-24 14:00:13.403474844 +0000 UTC m=+828.691232137" Nov 24 14:00:14 crc kubenswrapper[4970]: I1124 14:00:14.252152 4970 generic.go:334] "Generic (PLEG): container finished" podID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerID="069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65" exitCode=0 Nov 24 14:00:14 crc kubenswrapper[4970]: I1124 14:00:14.252290 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerDied","Data":"069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65"} Nov 24 14:00:14 crc kubenswrapper[4970]: I1124 14:00:14.252467 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerStarted","Data":"d9b48c3f292cb1b184192cede165de1a72cc9632c0abb9d043d4dc6dc6e8f962"} Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.330173 4970 generic.go:334] "Generic (PLEG): container finished" podID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerID="196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5" exitCode=0 Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.330270 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerDied","Data":"196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5"} Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.686175 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-86d796d84d-vs899" Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.705380 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6fdc856c5d-g8kmd" Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.734342 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5bdf4f7f7f-hdflg" Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.753259 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6dc664666c-t4pqg" Nov 24 14:00:22 crc kubenswrapper[4970]: I1124 14:00:22.964674 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7798859c74-k5b5j" Nov 24 14:00:23 crc kubenswrapper[4970]: I1124 14:00:23.174278 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp" Nov 24 14:00:23 crc kubenswrapper[4970]: I1124 14:00:23.176729 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7cd4fb6f79-m9m2l" Nov 24 14:00:23 crc kubenswrapper[4970]: I1124 14:00:23.339951 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerStarted","Data":"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732"} Nov 24 14:00:23 crc kubenswrapper[4970]: I1124 14:00:23.367820 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xwjpn" podStartSLOduration=7.883012253 podStartE2EDuration="16.367791325s" podCreationTimestamp="2025-11-24 14:00:07 +0000 UTC" firstStartedPulling="2025-11-24 14:00:14.25641598 +0000 UTC m=+829.544173303" lastFinishedPulling="2025-11-24 14:00:22.741195082 +0000 UTC m=+838.028952375" observedRunningTime="2025-11-24 14:00:23.358217266 +0000 UTC m=+838.645974559" watchObservedRunningTime="2025-11-24 14:00:23.367791325 +0000 UTC m=+838.655548648" Nov 24 14:00:27 crc kubenswrapper[4970]: I1124 14:00:27.864865 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:27 crc kubenswrapper[4970]: I1124 14:00:27.865502 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:28 crc kubenswrapper[4970]: I1124 14:00:28.145296 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:28 crc kubenswrapper[4970]: I1124 14:00:28.420739 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:28 crc kubenswrapper[4970]: I1124 14:00:28.471415 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:30 crc kubenswrapper[4970]: I1124 14:00:30.401938 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xwjpn" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="registry-server" containerID="cri-o://a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732" gracePeriod=2 Nov 24 14:00:30 crc kubenswrapper[4970]: I1124 14:00:30.865290 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.061544 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content\") pod \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.061726 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities\") pod \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.061918 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vh4g\" (UniqueName: \"kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g\") pod \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\" (UID: \"f1075b33-2cfe-4606-aa6d-53337a31bd9d\") " Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.062525 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities" (OuterVolumeSpecName: "utilities") pod "f1075b33-2cfe-4606-aa6d-53337a31bd9d" (UID: "f1075b33-2cfe-4606-aa6d-53337a31bd9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.063722 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.067915 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g" (OuterVolumeSpecName: "kube-api-access-6vh4g") pod "f1075b33-2cfe-4606-aa6d-53337a31bd9d" (UID: "f1075b33-2cfe-4606-aa6d-53337a31bd9d"). InnerVolumeSpecName "kube-api-access-6vh4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.083410 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1075b33-2cfe-4606-aa6d-53337a31bd9d" (UID: "f1075b33-2cfe-4606-aa6d-53337a31bd9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.166272 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1075b33-2cfe-4606-aa6d-53337a31bd9d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.166328 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vh4g\" (UniqueName: \"kubernetes.io/projected/f1075b33-2cfe-4606-aa6d-53337a31bd9d-kube-api-access-6vh4g\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.415682 4970 generic.go:334] "Generic (PLEG): container finished" podID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerID="a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732" exitCode=0 Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.415753 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerDied","Data":"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732"} Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.415808 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xwjpn" event={"ID":"f1075b33-2cfe-4606-aa6d-53337a31bd9d","Type":"ContainerDied","Data":"d9b48c3f292cb1b184192cede165de1a72cc9632c0abb9d043d4dc6dc6e8f962"} Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.415813 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xwjpn" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.415836 4970 scope.go:117] "RemoveContainer" containerID="a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.446296 4970 scope.go:117] "RemoveContainer" containerID="196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.510833 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.511212 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xwjpn"] Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.511699 4970 scope.go:117] "RemoveContainer" containerID="069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.527791 4970 scope.go:117] "RemoveContainer" containerID="a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732" Nov 24 14:00:31 crc kubenswrapper[4970]: E1124 14:00:31.528287 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732\": container with ID starting with a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732 not found: ID does not exist" containerID="a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.528323 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732"} err="failed to get container status \"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732\": rpc error: code = NotFound desc = could not find container \"a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732\": container with ID starting with a7f3f218a0e85da3df51910b819fab8d2a46f0864576074ca8b73ab7b842d732 not found: ID does not exist" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.528343 4970 scope.go:117] "RemoveContainer" containerID="196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5" Nov 24 14:00:31 crc kubenswrapper[4970]: E1124 14:00:31.528600 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5\": container with ID starting with 196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5 not found: ID does not exist" containerID="196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.528637 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5"} err="failed to get container status \"196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5\": rpc error: code = NotFound desc = could not find container \"196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5\": container with ID starting with 196fb87f78fa4a1e224051444fe180cd2d2bd621befd5c510a438841d226fca5 not found: ID does not exist" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.528664 4970 scope.go:117] "RemoveContainer" containerID="069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65" Nov 24 14:00:31 crc kubenswrapper[4970]: E1124 14:00:31.529139 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65\": container with ID starting with 069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65 not found: ID does not exist" containerID="069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65" Nov 24 14:00:31 crc kubenswrapper[4970]: I1124 14:00:31.529162 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65"} err="failed to get container status \"069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65\": rpc error: code = NotFound desc = could not find container \"069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65\": container with ID starting with 069752f37334f1aa0d5029e7b825d49bd05083f4b77192f3d9b04aa5ff9f3a65 not found: ID does not exist" Nov 24 14:00:33 crc kubenswrapper[4970]: I1124 14:00:33.481969 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" path="/var/lib/kubelet/pods/f1075b33-2cfe-4606-aa6d-53337a31bd9d/volumes" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.978255 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:36 crc kubenswrapper[4970]: E1124 14:00:36.978951 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="extract-content" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.978969 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="extract-content" Nov 24 14:00:36 crc kubenswrapper[4970]: E1124 14:00:36.978994 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="registry-server" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.979001 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="registry-server" Nov 24 14:00:36 crc kubenswrapper[4970]: E1124 14:00:36.979041 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="extract-utilities" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.979051 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="extract-utilities" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.979216 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1075b33-2cfe-4606-aa6d-53337a31bd9d" containerName="registry-server" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.980727 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.983960 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.984809 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.985030 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 14:00:36 crc kubenswrapper[4970]: I1124 14:00:36.987386 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-s7rbp" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.000885 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.068392 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.070948 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.079051 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.089299 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.153987 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqqd\" (UniqueName: \"kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.154043 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.154071 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.154112 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.154162 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pthh\" (UniqueName: \"kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.255516 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.255632 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pthh\" (UniqueName: \"kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.255691 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqqd\" (UniqueName: \"kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.255737 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.255777 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.256364 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.256539 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.256827 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.276608 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqqd\" (UniqueName: \"kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd\") pod \"dnsmasq-dns-6584b49599-v947x\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.280216 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pthh\" (UniqueName: \"kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh\") pod \"dnsmasq-dns-7bdd77c89-2trtj\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.300798 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.424683 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.751116 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.756117 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:00:37 crc kubenswrapper[4970]: I1124 14:00:37.852010 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:37 crc kubenswrapper[4970]: W1124 14:00:37.859120 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc84189_d32d_42b0_b863_efd1ba8fc55d.slice/crio-ff02c05b3cac71001ff641bdecf4eb6e50e3f34ca4e1acbb746f3e940a3e7950 WatchSource:0}: Error finding container ff02c05b3cac71001ff641bdecf4eb6e50e3f34ca4e1acbb746f3e940a3e7950: Status 404 returned error can't find the container with id ff02c05b3cac71001ff641bdecf4eb6e50e3f34ca4e1acbb746f3e940a3e7950 Nov 24 14:00:38 crc kubenswrapper[4970]: I1124 14:00:38.468407 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" event={"ID":"06be1c38-04e6-4743-a07e-00c0a6f5bf45","Type":"ContainerStarted","Data":"aa0ad7c7ba08af0b7e29c49f0a45f02fcb2bd2e39ed53d34d53f3da58f975eb9"} Nov 24 14:00:38 crc kubenswrapper[4970]: I1124 14:00:38.472268 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-v947x" event={"ID":"5dc84189-d32d-42b0-b863-efd1ba8fc55d","Type":"ContainerStarted","Data":"ff02c05b3cac71001ff641bdecf4eb6e50e3f34ca4e1acbb746f3e940a3e7950"} Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.404369 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.424926 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.426353 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.436372 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.510618 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.510683 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pxmg\" (UniqueName: \"kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.510752 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.612616 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.612695 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pxmg\" (UniqueName: \"kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.612743 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.613891 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.613910 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.643826 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pxmg\" (UniqueName: \"kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg\") pod \"dnsmasq-dns-7c6d9948dc-sfk57\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.718339 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.737944 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.739085 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.745146 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.755066 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.816997 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.817128 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k997\" (UniqueName: \"kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.817161 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.918740 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.918856 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k997\" (UniqueName: \"kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.918880 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.919746 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.919813 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:40 crc kubenswrapper[4970]: I1124 14:00:40.961464 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k997\" (UniqueName: \"kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997\") pod \"dnsmasq-dns-6486446b9f-cqdjs\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.055819 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.587329 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.588604 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593035 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593157 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593249 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593377 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593476 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593569 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fwx5r" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.593638 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.607406 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.729267 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730255 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndd2p\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730289 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730312 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730341 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730420 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730464 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730510 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730564 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730686 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.730714 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836505 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836552 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836589 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836605 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836646 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836676 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndd2p\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836697 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836713 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836734 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836760 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836778 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.836971 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.837521 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.838087 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.838255 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.837301 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.839357 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.845714 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.847943 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.848350 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.848674 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.857829 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndd2p\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.861760 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " pod="openstack/rabbitmq-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.903250 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.904383 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.906428 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.906680 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.906830 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.906972 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.907083 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.907872 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.907967 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mz7hf" Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.925916 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:00:41 crc kubenswrapper[4970]: I1124 14:00:41.927533 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.044944 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045026 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045059 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045104 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045149 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq6ff\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045192 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045228 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045251 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045272 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.045295 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.146975 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147048 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147107 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq6ff\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147168 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147212 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147241 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147264 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147289 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147314 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147346 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.147368 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.148524 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.149268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.149825 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.149891 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.150417 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.151071 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.152927 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.154506 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.167416 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.168958 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.177972 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq6ff\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.182061 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:42 crc kubenswrapper[4970]: I1124 14:00:42.248902 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.193277 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.195371 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.205079 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.205817 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-t6zt2" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.205851 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.205907 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.211099 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.211747 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365490 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365548 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365642 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365674 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-default\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365702 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365745 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365779 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kolla-config\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.365803 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hjg\" (UniqueName: \"kubernetes.io/projected/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kube-api-access-q5hjg\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469384 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469431 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469466 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469499 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-default\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469525 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469567 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469622 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kolla-config\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469650 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hjg\" (UniqueName: \"kubernetes.io/projected/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kube-api-access-q5hjg\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469840 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.469931 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.470673 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-config-data-default\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.471114 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.471631 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kolla-config\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.473902 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.478085 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b6f151-1b32-4346-aabb-38ce2d89f50f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.497811 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hjg\" (UniqueName: \"kubernetes.io/projected/f0b6f151-1b32-4346-aabb-38ce2d89f50f-kube-api-access-q5hjg\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.500987 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"f0b6f151-1b32-4346-aabb-38ce2d89f50f\") " pod="openstack/openstack-galera-0" Nov 24 14:00:43 crc kubenswrapper[4970]: I1124 14:00:43.524618 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.278298 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.604836 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.607790 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.607882 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.610478 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-h92rs" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.616374 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.616633 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.619418 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687225 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687298 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687320 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kube-api-access-rzgfr\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687336 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687358 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687409 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687424 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.687447 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788590 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788685 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788722 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kube-api-access-rzgfr\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788751 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788785 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788858 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788884 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.788916 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.789453 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.790073 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.790215 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.790388 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.791373 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.794328 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.804570 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzgfr\" (UniqueName: \"kubernetes.io/projected/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-kube-api-access-rzgfr\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.818082 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/69fa6c84-911e-41f0-904d-1bd1b08d3b0c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.826660 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.827653 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.831983 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-8l58l" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.832180 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.832291 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.837782 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.877270 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"69fa6c84-911e-41f0-904d-1bd1b08d3b0c\") " pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.890557 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-kolla-config\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.890649 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.890678 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5zlt\" (UniqueName: \"kubernetes.io/projected/40d13307-a86f-453f-a3c8-05f936d03e73-kube-api-access-v5zlt\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.890735 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.890766 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-config-data\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.935365 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.991882 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-kolla-config\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.991948 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.991972 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5zlt\" (UniqueName: \"kubernetes.io/projected/40d13307-a86f-453f-a3c8-05f936d03e73-kube-api-access-v5zlt\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.992013 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.992036 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-config-data\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.992860 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-config-data\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.993495 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/40d13307-a86f-453f-a3c8-05f936d03e73-kolla-config\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:44 crc kubenswrapper[4970]: I1124 14:00:44.996512 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-combined-ca-bundle\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:45 crc kubenswrapper[4970]: I1124 14:00:45.004271 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/40d13307-a86f-453f-a3c8-05f936d03e73-memcached-tls-certs\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:45 crc kubenswrapper[4970]: I1124 14:00:45.014264 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5zlt\" (UniqueName: \"kubernetes.io/projected/40d13307-a86f-453f-a3c8-05f936d03e73-kube-api-access-v5zlt\") pod \"memcached-0\" (UID: \"40d13307-a86f-453f-a3c8-05f936d03e73\") " pod="openstack/memcached-0" Nov 24 14:00:45 crc kubenswrapper[4970]: I1124 14:00:45.220930 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.475472 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.479906 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.483216 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zc5s8" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.494382 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.620418 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpxpw\" (UniqueName: \"kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw\") pod \"kube-state-metrics-0\" (UID: \"26f07678-0be1-4b4d-9754-8a8c54b8d82a\") " pod="openstack/kube-state-metrics-0" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.721462 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpxpw\" (UniqueName: \"kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw\") pod \"kube-state-metrics-0\" (UID: \"26f07678-0be1-4b4d-9754-8a8c54b8d82a\") " pod="openstack/kube-state-metrics-0" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.761641 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpxpw\" (UniqueName: \"kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw\") pod \"kube-state-metrics-0\" (UID: \"26f07678-0be1-4b4d-9754-8a8c54b8d82a\") " pod="openstack/kube-state-metrics-0" Nov 24 14:00:46 crc kubenswrapper[4970]: I1124 14:00:46.808462 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:00:48 crc kubenswrapper[4970]: I1124 14:00:48.068279 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:00:48 crc kubenswrapper[4970]: I1124 14:00:48.582273 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" event={"ID":"fef86e1d-5b44-4213-9adc-2da9c188f733","Type":"ContainerStarted","Data":"d814d8f1c01ebc28332dc18274a2e3f7fea904ba655da467c416c9743ba3da55"} Nov 24 14:00:50 crc kubenswrapper[4970]: I1124 14:00:50.961011 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-78b67"] Nov 24 14:00:50 crc kubenswrapper[4970]: I1124 14:00:50.962495 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67" Nov 24 14:00:50 crc kubenswrapper[4970]: I1124 14:00:50.972631 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67"] Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.002986 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.003024 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6wjfh" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.003340 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004676 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb1eaff5-0428-4ab0-ae3b-21f192c95279-scripts\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004725 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-ovn-controller-tls-certs\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004766 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn9sn\" (UniqueName: \"kubernetes.io/projected/bb1eaff5-0428-4ab0-ae3b-21f192c95279-kube-api-access-cn9sn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004864 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004913 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004956 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-log-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.004985 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-combined-ca-bundle\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.045781 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-r89nk"] Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.047449 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.052348 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-r89nk"] Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106656 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn9sn\" (UniqueName: \"kubernetes.io/projected/bb1eaff5-0428-4ab0-ae3b-21f192c95279-kube-api-access-cn9sn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106736 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106763 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-lib\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106787 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-run\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106806 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106829 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdt5l\" (UniqueName: \"kubernetes.io/projected/c5806c43-906c-4790-82e3-511bd817eb88-kube-api-access-bdt5l\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106864 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-log-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.106883 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-combined-ca-bundle\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107015 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-log\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107063 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5806c43-906c-4790-82e3-511bd817eb88-scripts\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107093 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb1eaff5-0428-4ab0-ae3b-21f192c95279-scripts\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107109 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-etc-ovs\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107133 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-ovn-controller-tls-certs\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107411 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107456 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-log-ovn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.107512 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb1eaff5-0428-4ab0-ae3b-21f192c95279-var-run\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.110223 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb1eaff5-0428-4ab0-ae3b-21f192c95279-scripts\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.114138 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-combined-ca-bundle\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.117254 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb1eaff5-0428-4ab0-ae3b-21f192c95279-ovn-controller-tls-certs\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.126045 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn9sn\" (UniqueName: \"kubernetes.io/projected/bb1eaff5-0428-4ab0-ae3b-21f192c95279-kube-api-access-cn9sn\") pod \"ovn-controller-78b67\" (UID: \"bb1eaff5-0428-4ab0-ae3b-21f192c95279\") " pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.188167 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.201561 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.201744 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.206434 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-jqdmz" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.206730 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.206802 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.207028 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.208863 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-log\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.208896 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5806c43-906c-4790-82e3-511bd817eb88-scripts\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.208936 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-etc-ovs\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.208983 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-lib\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209010 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-run\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209045 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdt5l\" (UniqueName: \"kubernetes.io/projected/c5806c43-906c-4790-82e3-511bd817eb88-kube-api-access-bdt5l\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209559 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-log\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209731 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-lib\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209857 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-etc-ovs\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.209906 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c5806c43-906c-4790-82e3-511bd817eb88-var-run\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.211916 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.213884 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5806c43-906c-4790-82e3-511bd817eb88-scripts\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.231211 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdt5l\" (UniqueName: \"kubernetes.io/projected/c5806c43-906c-4790-82e3-511bd817eb88-kube-api-access-bdt5l\") pod \"ovn-controller-ovs-r89nk\" (UID: \"c5806c43-906c-4790-82e3-511bd817eb88\") " pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.310644 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311235 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311284 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-config\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311403 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311444 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs97k\" (UniqueName: \"kubernetes.io/projected/90130f89-b1ad-43b9-9ab4-296530171bd6-kube-api-access-fs97k\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311503 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311704 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.311772 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.321916 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.366248 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413136 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413199 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413225 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413242 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413268 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-config\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413300 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413315 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs97k\" (UniqueName: \"kubernetes.io/projected/90130f89-b1ad-43b9-9ab4-296530171bd6-kube-api-access-fs97k\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413338 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.413783 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.414101 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.414397 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-config\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.415295 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90130f89-b1ad-43b9-9ab4-296530171bd6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.419164 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.426984 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.432031 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90130f89-b1ad-43b9-9ab4-296530171bd6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.432401 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs97k\" (UniqueName: \"kubernetes.io/projected/90130f89-b1ad-43b9-9ab4-296530171bd6-kube-api-access-fs97k\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.443372 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90130f89-b1ad-43b9-9ab4-296530171bd6\") " pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:51 crc kubenswrapper[4970]: I1124 14:00:51.532019 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.053839 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:00:53 crc kubenswrapper[4970]: W1124 14:00:53.653647 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74a8ae4a_b366_490b_aced_d4cad00248c0.slice/crio-5b9ae162b1bdb365fef44b1e019f69922355c711a92fb7ef68d7a9add242e9ee WatchSource:0}: Error finding container 5b9ae162b1bdb365fef44b1e019f69922355c711a92fb7ef68d7a9add242e9ee: Status 404 returned error can't find the container with id 5b9ae162b1bdb365fef44b1e019f69922355c711a92fb7ef68d7a9add242e9ee Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.661915 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerStarted","Data":"81aec99ec46ae49b39341c917c65896f82d894ee50d19f5e8d7bd3ae27843d2f"} Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.718503 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.718745 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2pthh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-2trtj_openstack(06be1c38-04e6-4743-a07e-00c0a6f5bf45): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.719966 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" podUID="06be1c38-04e6-4743-a07e-00c0a6f5bf45" Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.876356 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.876504 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqqqd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-v947x_openstack(5dc84189-d32d-42b0-b863-efd1ba8fc55d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:00:53 crc kubenswrapper[4970]: E1124 14:00:53.877711 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-v947x" podUID="5dc84189-d32d-42b0-b863-efd1ba8fc55d" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.899069 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.900874 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.903120 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-v9ssj" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.905851 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.906115 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.906977 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 14:00:53 crc kubenswrapper[4970]: I1124 14:00:53.915568 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071632 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071691 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071707 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071761 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071776 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071804 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-config\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.071968 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qsjj\" (UniqueName: \"kubernetes.io/projected/4910fec1-46d5-419d-baf9-3fa44a49b1d5-kube-api-access-8qsjj\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.072043 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.177938 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qsjj\" (UniqueName: \"kubernetes.io/projected/4910fec1-46d5-419d-baf9-3fa44a49b1d5-kube-api-access-8qsjj\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.177986 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178014 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178047 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178062 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178112 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178127 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178154 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-config\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.178480 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.179231 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.182894 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.183140 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4910fec1-46d5-419d-baf9-3fa44a49b1d5-config\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.183640 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.183645 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.183646 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4910fec1-46d5-419d-baf9-3fa44a49b1d5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.193408 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qsjj\" (UniqueName: \"kubernetes.io/projected/4910fec1-46d5-419d-baf9-3fa44a49b1d5-kube-api-access-8qsjj\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.200779 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4910fec1-46d5-419d-baf9-3fa44a49b1d5\") " pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.230821 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.391939 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.402246 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 14:00:54 crc kubenswrapper[4970]: W1124 14:00:54.411187 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0b6f151_1b32_4346_aabb_38ce2d89f50f.slice/crio-a9895f65c569895bb031c5a41972c37736590d258fd63001854069aba1b9c584 WatchSource:0}: Error finding container a9895f65c569895bb031c5a41972c37736590d258fd63001854069aba1b9c584: Status 404 returned error can't find the container with id a9895f65c569895bb031c5a41972c37736590d258fd63001854069aba1b9c584 Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.531792 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 14:00:54 crc kubenswrapper[4970]: W1124 14:00:54.537306 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69fa6c84_911e_41f0_904d_1bd1b08d3b0c.slice/crio-e57966ef4cfcadad84ad952281d42a67b80d20634144a4fd3a3ae302beba0f7f WatchSource:0}: Error finding container e57966ef4cfcadad84ad952281d42a67b80d20634144a4fd3a3ae302beba0f7f: Status 404 returned error can't find the container with id e57966ef4cfcadad84ad952281d42a67b80d20634144a4fd3a3ae302beba0f7f Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.556941 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.610695 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67"] Nov 24 14:00:54 crc kubenswrapper[4970]: W1124 14:00:54.614756 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb1eaff5_0428_4ab0_ae3b_21f192c95279.slice/crio-759aafb22da00930459e536fff93f423c4f7d1decc52e58c46ceedea44398082 WatchSource:0}: Error finding container 759aafb22da00930459e536fff93f423c4f7d1decc52e58c46ceedea44398082: Status 404 returned error can't find the container with id 759aafb22da00930459e536fff93f423c4f7d1decc52e58c46ceedea44398082 Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.677779 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerStarted","Data":"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.677847 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerStarted","Data":"7e73c8df5d24f05cb86471ec532102e5baa2f31bdaf7c80bb0254c81fa7b5d28"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.679375 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67" event={"ID":"bb1eaff5-0428-4ab0-ae3b-21f192c95279","Type":"ContainerStarted","Data":"759aafb22da00930459e536fff93f423c4f7d1decc52e58c46ceedea44398082"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.682783 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerStarted","Data":"5b9ae162b1bdb365fef44b1e019f69922355c711a92fb7ef68d7a9add242e9ee"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.684113 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26f07678-0be1-4b4d-9754-8a8c54b8d82a","Type":"ContainerStarted","Data":"ebff66a78c61b7d9033174f06a5f6032ca7f27a723d91064af49be3b55b1766a"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.685387 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"69fa6c84-911e-41f0-904d-1bd1b08d3b0c","Type":"ContainerStarted","Data":"e57966ef4cfcadad84ad952281d42a67b80d20634144a4fd3a3ae302beba0f7f"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.687403 4970 generic.go:334] "Generic (PLEG): container finished" podID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerID="1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37" exitCode=0 Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.687468 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" event={"ID":"fef86e1d-5b44-4213-9adc-2da9c188f733","Type":"ContainerDied","Data":"1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.692691 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f0b6f151-1b32-4346-aabb-38ce2d89f50f","Type":"ContainerStarted","Data":"a9895f65c569895bb031c5a41972c37736590d258fd63001854069aba1b9c584"} Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.734027 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-r89nk"] Nov 24 14:00:54 crc kubenswrapper[4970]: W1124 14:00:54.784539 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5806c43_906c_4790_82e3_511bd817eb88.slice/crio-7989e107ed0478ba695cfe3657f850c88e7657362bc3908b6305136b13c80929 WatchSource:0}: Error finding container 7989e107ed0478ba695cfe3657f850c88e7657362bc3908b6305136b13c80929: Status 404 returned error can't find the container with id 7989e107ed0478ba695cfe3657f850c88e7657362bc3908b6305136b13c80929 Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.825650 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 14:00:54 crc kubenswrapper[4970]: W1124 14:00:54.850320 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40d13307_a86f_453f_a3c8_05f936d03e73.slice/crio-be1566cf5b3399b680a89469d2afa28000f76f0c479ad61905192d28383f2109 WatchSource:0}: Error finding container be1566cf5b3399b680a89469d2afa28000f76f0c479ad61905192d28383f2109: Status 404 returned error can't find the container with id be1566cf5b3399b680a89469d2afa28000f76f0c479ad61905192d28383f2109 Nov 24 14:00:54 crc kubenswrapper[4970]: I1124 14:00:54.963743 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.085403 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.138395 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.193004 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqqqd\" (UniqueName: \"kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd\") pod \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.193087 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc\") pod \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.193182 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config\") pod \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\" (UID: \"5dc84189-d32d-42b0-b863-efd1ba8fc55d\") " Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.193604 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5dc84189-d32d-42b0-b863-efd1ba8fc55d" (UID: "5dc84189-d32d-42b0-b863-efd1ba8fc55d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.193977 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config" (OuterVolumeSpecName: "config") pod "5dc84189-d32d-42b0-b863-efd1ba8fc55d" (UID: "5dc84189-d32d-42b0-b863-efd1ba8fc55d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.198943 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd" (OuterVolumeSpecName: "kube-api-access-fqqqd") pod "5dc84189-d32d-42b0-b863-efd1ba8fc55d" (UID: "5dc84189-d32d-42b0-b863-efd1ba8fc55d"). InnerVolumeSpecName "kube-api-access-fqqqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.294697 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pthh\" (UniqueName: \"kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh\") pod \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.294761 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config\") pod \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\" (UID: \"06be1c38-04e6-4743-a07e-00c0a6f5bf45\") " Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.295107 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqqqd\" (UniqueName: \"kubernetes.io/projected/5dc84189-d32d-42b0-b863-efd1ba8fc55d-kube-api-access-fqqqd\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.295127 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.295136 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc84189-d32d-42b0-b863-efd1ba8fc55d-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.295235 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config" (OuterVolumeSpecName: "config") pod "06be1c38-04e6-4743-a07e-00c0a6f5bf45" (UID: "06be1c38-04e6-4743-a07e-00c0a6f5bf45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.297798 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh" (OuterVolumeSpecName: "kube-api-access-2pthh") pod "06be1c38-04e6-4743-a07e-00c0a6f5bf45" (UID: "06be1c38-04e6-4743-a07e-00c0a6f5bf45"). InnerVolumeSpecName "kube-api-access-2pthh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.396998 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pthh\" (UniqueName: \"kubernetes.io/projected/06be1c38-04e6-4743-a07e-00c0a6f5bf45-kube-api-access-2pthh\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.397034 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06be1c38-04e6-4743-a07e-00c0a6f5bf45-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.615967 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 14:00:55 crc kubenswrapper[4970]: W1124 14:00:55.626558 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90130f89_b1ad_43b9_9ab4_296530171bd6.slice/crio-9456c5cd1d94a1811d68bc79b7fe2bbd20b9ed0d625c9a50433818ce62c81527 WatchSource:0}: Error finding container 9456c5cd1d94a1811d68bc79b7fe2bbd20b9ed0d625c9a50433818ce62c81527: Status 404 returned error can't find the container with id 9456c5cd1d94a1811d68bc79b7fe2bbd20b9ed0d625c9a50433818ce62c81527 Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.703591 4970 generic.go:334] "Generic (PLEG): container finished" podID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerID="446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267" exitCode=0 Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.703658 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerDied","Data":"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.707873 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4910fec1-46d5-419d-baf9-3fa44a49b1d5","Type":"ContainerStarted","Data":"d3d5216110dcc582db09327f8847a9c787184ec4c32476a8507ed3a22d0b88ee"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.709170 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90130f89-b1ad-43b9-9ab4-296530171bd6","Type":"ContainerStarted","Data":"9456c5cd1d94a1811d68bc79b7fe2bbd20b9ed0d625c9a50433818ce62c81527"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.710349 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40d13307-a86f-453f-a3c8-05f936d03e73","Type":"ContainerStarted","Data":"be1566cf5b3399b680a89469d2afa28000f76f0c479ad61905192d28383f2109"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.711399 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r89nk" event={"ID":"c5806c43-906c-4790-82e3-511bd817eb88","Type":"ContainerStarted","Data":"7989e107ed0478ba695cfe3657f850c88e7657362bc3908b6305136b13c80929"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.712380 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-v947x" event={"ID":"5dc84189-d32d-42b0-b863-efd1ba8fc55d","Type":"ContainerDied","Data":"ff02c05b3cac71001ff641bdecf4eb6e50e3f34ca4e1acbb746f3e940a3e7950"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.712432 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v947x" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.714327 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" event={"ID":"fef86e1d-5b44-4213-9adc-2da9c188f733","Type":"ContainerStarted","Data":"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.714456 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.716535 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" event={"ID":"06be1c38-04e6-4743-a07e-00c0a6f5bf45","Type":"ContainerDied","Data":"aa0ad7c7ba08af0b7e29c49f0a45f02fcb2bd2e39ed53d34d53f3da58f975eb9"} Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.716541 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-2trtj" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.751191 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" podStartSLOduration=9.296839695 podStartE2EDuration="15.751166833s" podCreationTimestamp="2025-11-24 14:00:40 +0000 UTC" firstStartedPulling="2025-11-24 14:00:47.60542985 +0000 UTC m=+862.893187153" lastFinishedPulling="2025-11-24 14:00:54.059756998 +0000 UTC m=+869.347514291" observedRunningTime="2025-11-24 14:00:55.742165258 +0000 UTC m=+871.029922551" watchObservedRunningTime="2025-11-24 14:00:55.751166833 +0000 UTC m=+871.038924126" Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.822333 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.834359 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v947x"] Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.851769 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:55 crc kubenswrapper[4970]: I1124 14:00:55.856984 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-2trtj"] Nov 24 14:00:56 crc kubenswrapper[4970]: I1124 14:00:56.724524 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerStarted","Data":"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd"} Nov 24 14:00:56 crc kubenswrapper[4970]: I1124 14:00:56.745459 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" podStartSLOduration=16.745439329 podStartE2EDuration="16.745439329s" podCreationTimestamp="2025-11-24 14:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:00:56.738448027 +0000 UTC m=+872.026205330" watchObservedRunningTime="2025-11-24 14:00:56.745439329 +0000 UTC m=+872.033196622" Nov 24 14:00:57 crc kubenswrapper[4970]: I1124 14:00:57.481304 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06be1c38-04e6-4743-a07e-00c0a6f5bf45" path="/var/lib/kubelet/pods/06be1c38-04e6-4743-a07e-00c0a6f5bf45/volumes" Nov 24 14:00:57 crc kubenswrapper[4970]: I1124 14:00:57.481910 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc84189-d32d-42b0-b863-efd1ba8fc55d" path="/var/lib/kubelet/pods/5dc84189-d32d-42b0-b863-efd1ba8fc55d/volumes" Nov 24 14:00:57 crc kubenswrapper[4970]: I1124 14:00:57.730756 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:01:00 crc kubenswrapper[4970]: I1124 14:01:00.746782 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:01:01 crc kubenswrapper[4970]: I1124 14:01:01.058627 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:01:01 crc kubenswrapper[4970]: I1124 14:01:01.112026 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:01:01 crc kubenswrapper[4970]: I1124 14:01:01.759790 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="dnsmasq-dns" containerID="cri-o://cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd" gracePeriod=10 Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.264390 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.316334 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc\") pod \"644eb975-00bf-42a7-ab6c-332544ea31d6\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.316477 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config\") pod \"644eb975-00bf-42a7-ab6c-332544ea31d6\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.316534 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pxmg\" (UniqueName: \"kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg\") pod \"644eb975-00bf-42a7-ab6c-332544ea31d6\" (UID: \"644eb975-00bf-42a7-ab6c-332544ea31d6\") " Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.346753 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg" (OuterVolumeSpecName: "kube-api-access-9pxmg") pod "644eb975-00bf-42a7-ab6c-332544ea31d6" (UID: "644eb975-00bf-42a7-ab6c-332544ea31d6"). InnerVolumeSpecName "kube-api-access-9pxmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.371027 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "644eb975-00bf-42a7-ab6c-332544ea31d6" (UID: "644eb975-00bf-42a7-ab6c-332544ea31d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.383878 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config" (OuterVolumeSpecName: "config") pod "644eb975-00bf-42a7-ab6c-332544ea31d6" (UID: "644eb975-00bf-42a7-ab6c-332544ea31d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.418155 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.418191 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/644eb975-00bf-42a7-ab6c-332544ea31d6-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.418202 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pxmg\" (UniqueName: \"kubernetes.io/projected/644eb975-00bf-42a7-ab6c-332544ea31d6-kube-api-access-9pxmg\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.772353 4970 generic.go:334] "Generic (PLEG): container finished" podID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerID="cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd" exitCode=0 Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.772415 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerDied","Data":"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd"} Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.772731 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" event={"ID":"644eb975-00bf-42a7-ab6c-332544ea31d6","Type":"ContainerDied","Data":"7e73c8df5d24f05cb86471ec532102e5baa2f31bdaf7c80bb0254c81fa7b5d28"} Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.772428 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-sfk57" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.772779 4970 scope.go:117] "RemoveContainer" containerID="cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.804401 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.810787 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-sfk57"] Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.893646 4970 scope.go:117] "RemoveContainer" containerID="446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.920172 4970 scope.go:117] "RemoveContainer" containerID="cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd" Nov 24 14:01:02 crc kubenswrapper[4970]: E1124 14:01:02.920858 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd\": container with ID starting with cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd not found: ID does not exist" containerID="cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.920914 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd"} err="failed to get container status \"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd\": rpc error: code = NotFound desc = could not find container \"cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd\": container with ID starting with cfee8496fa990366726993e3a229e0063d2018a77af78604351acb43e05014cd not found: ID does not exist" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.920945 4970 scope.go:117] "RemoveContainer" containerID="446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267" Nov 24 14:01:02 crc kubenswrapper[4970]: E1124 14:01:02.921415 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267\": container with ID starting with 446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267 not found: ID does not exist" containerID="446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267" Nov 24 14:01:02 crc kubenswrapper[4970]: I1124 14:01:02.921460 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267"} err="failed to get container status \"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267\": rpc error: code = NotFound desc = could not find container \"446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267\": container with ID starting with 446a52ad9be7f0670fc87cfcb2164913c126bdff520a944e9364fb7057dd3267 not found: ID does not exist" Nov 24 14:01:03 crc kubenswrapper[4970]: I1124 14:01:03.485671 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" path="/var/lib/kubelet/pods/644eb975-00bf-42a7-ab6c-332544ea31d6/volumes" Nov 24 14:01:03 crc kubenswrapper[4970]: I1124 14:01:03.789641 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"69fa6c84-911e-41f0-904d-1bd1b08d3b0c","Type":"ContainerStarted","Data":"fac65660cd60dff37d5a62bcebd54e995075ad4859a0003fc444546c0c091b9c"} Nov 24 14:01:03 crc kubenswrapper[4970]: I1124 14:01:03.791922 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerStarted","Data":"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974"} Nov 24 14:01:03 crc kubenswrapper[4970]: I1124 14:01:03.798115 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f0b6f151-1b32-4346-aabb-38ce2d89f50f","Type":"ContainerStarted","Data":"385e297e76b29c580fa5f89a11d729f9ce8bfc438caa6249bdf2f2f38bea8b4c"} Nov 24 14:01:03 crc kubenswrapper[4970]: I1124 14:01:03.805216 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90130f89-b1ad-43b9-9ab4-296530171bd6","Type":"ContainerStarted","Data":"f18adab042ff7bcec84dc89e8204ac9eb98e66a0a97175f929aa4dad1817c424"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.813638 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerStarted","Data":"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.815807 4970 generic.go:334] "Generic (PLEG): container finished" podID="c5806c43-906c-4790-82e3-511bd817eb88" containerID="ec50e72ffa9d8e56748411e9d3181252554a4d91d479fd3f0928326faa4e770a" exitCode=0 Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.815849 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r89nk" event={"ID":"c5806c43-906c-4790-82e3-511bd817eb88","Type":"ContainerDied","Data":"ec50e72ffa9d8e56748411e9d3181252554a4d91d479fd3f0928326faa4e770a"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.817755 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26f07678-0be1-4b4d-9754-8a8c54b8d82a","Type":"ContainerStarted","Data":"d8eab8f64df417c0162763f8b1b084f62121581fc02b1f4a709d3d953f7a7ce8"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.817896 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.819210 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67" event={"ID":"bb1eaff5-0428-4ab0-ae3b-21f192c95279","Type":"ContainerStarted","Data":"fdfcdbc6c36a253fd82c30e73cdbafae9cd3e864e7abc74c76df355b645d83e9"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.819676 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-78b67" Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.822370 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4910fec1-46d5-419d-baf9-3fa44a49b1d5","Type":"ContainerStarted","Data":"d3b6e6605a79fbf901e1a4649554fac06498085b558d8e1db48a4a073c621cb0"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.826307 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"40d13307-a86f-453f-a3c8-05f936d03e73","Type":"ContainerStarted","Data":"ecaba010fa3a031ca0635ee85841e339b8a172e5a811c0b3560a6362837f2e8b"} Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.863743 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-78b67" podStartSLOduration=7.506277096 podStartE2EDuration="14.863728417s" podCreationTimestamp="2025-11-24 14:00:50 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.619287284 +0000 UTC m=+869.907044577" lastFinishedPulling="2025-11-24 14:01:01.976738595 +0000 UTC m=+877.264495898" observedRunningTime="2025-11-24 14:01:04.863153693 +0000 UTC m=+880.150910986" watchObservedRunningTime="2025-11-24 14:01:04.863728417 +0000 UTC m=+880.151485710" Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.902283 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.779936603 podStartE2EDuration="20.902264181s" podCreationTimestamp="2025-11-24 14:00:44 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.854695904 +0000 UTC m=+870.142453197" lastFinishedPulling="2025-11-24 14:01:01.977023482 +0000 UTC m=+877.264780775" observedRunningTime="2025-11-24 14:01:04.90222647 +0000 UTC m=+880.189983763" watchObservedRunningTime="2025-11-24 14:01:04.902264181 +0000 UTC m=+880.190021474" Nov 24 14:01:04 crc kubenswrapper[4970]: I1124 14:01:04.918783 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.553575634 podStartE2EDuration="18.91876721s" podCreationTimestamp="2025-11-24 14:00:46 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.580423063 +0000 UTC m=+869.868180356" lastFinishedPulling="2025-11-24 14:01:02.945614639 +0000 UTC m=+878.233371932" observedRunningTime="2025-11-24 14:01:04.91724303 +0000 UTC m=+880.205000333" watchObservedRunningTime="2025-11-24 14:01:04.91876721 +0000 UTC m=+880.206524503" Nov 24 14:01:05 crc kubenswrapper[4970]: I1124 14:01:05.222087 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 14:01:05 crc kubenswrapper[4970]: I1124 14:01:05.835324 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r89nk" event={"ID":"c5806c43-906c-4790-82e3-511bd817eb88","Type":"ContainerStarted","Data":"66f15f99f973a8b929fb9267decc9ae4230ccfcd95e0ed20166e381b9f0a50d8"} Nov 24 14:01:05 crc kubenswrapper[4970]: I1124 14:01:05.835720 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r89nk" event={"ID":"c5806c43-906c-4790-82e3-511bd817eb88","Type":"ContainerStarted","Data":"678c30cde7050591c959e205e361c77d04fd365389c92c3e713a2b22380b3cf0"} Nov 24 14:01:05 crc kubenswrapper[4970]: I1124 14:01:05.836217 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:01:05 crc kubenswrapper[4970]: I1124 14:01:05.860866 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-r89nk" podStartSLOduration=8.096262332 podStartE2EDuration="14.860839647s" podCreationTimestamp="2025-11-24 14:00:51 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.786927049 +0000 UTC m=+870.074684342" lastFinishedPulling="2025-11-24 14:01:01.551504364 +0000 UTC m=+876.839261657" observedRunningTime="2025-11-24 14:01:05.852735966 +0000 UTC m=+881.140493259" watchObservedRunningTime="2025-11-24 14:01:05.860839647 +0000 UTC m=+881.148596940" Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.367358 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.843867 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4910fec1-46d5-419d-baf9-3fa44a49b1d5","Type":"ContainerStarted","Data":"64d22584baedb5316c1ef21a9873829e39922feb5d448e2c658c3f7ff6bf1748"} Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.846632 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90130f89-b1ad-43b9-9ab4-296530171bd6","Type":"ContainerStarted","Data":"923cfd582fdd6a4758d7032621d34fa7a63673f6046670517a606a5f0acdeba6"} Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.848346 4970 generic.go:334] "Generic (PLEG): container finished" podID="f0b6f151-1b32-4346-aabb-38ce2d89f50f" containerID="385e297e76b29c580fa5f89a11d729f9ce8bfc438caa6249bdf2f2f38bea8b4c" exitCode=0 Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.848530 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f0b6f151-1b32-4346-aabb-38ce2d89f50f","Type":"ContainerDied","Data":"385e297e76b29c580fa5f89a11d729f9ce8bfc438caa6249bdf2f2f38bea8b4c"} Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.870557 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.319395662 podStartE2EDuration="14.870532964s" podCreationTimestamp="2025-11-24 14:00:52 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.991128726 +0000 UTC m=+870.278886019" lastFinishedPulling="2025-11-24 14:01:06.542266028 +0000 UTC m=+881.830023321" observedRunningTime="2025-11-24 14:01:06.863618004 +0000 UTC m=+882.151375297" watchObservedRunningTime="2025-11-24 14:01:06.870532964 +0000 UTC m=+882.158290267" Nov 24 14:01:06 crc kubenswrapper[4970]: I1124 14:01:06.886181 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.993889333 podStartE2EDuration="16.886151711s" podCreationTimestamp="2025-11-24 14:00:50 +0000 UTC" firstStartedPulling="2025-11-24 14:00:55.631852157 +0000 UTC m=+870.919609450" lastFinishedPulling="2025-11-24 14:01:06.524114535 +0000 UTC m=+881.811871828" observedRunningTime="2025-11-24 14:01:06.884332023 +0000 UTC m=+882.172089316" watchObservedRunningTime="2025-11-24 14:01:06.886151711 +0000 UTC m=+882.173909044" Nov 24 14:01:07 crc kubenswrapper[4970]: I1124 14:01:07.860379 4970 generic.go:334] "Generic (PLEG): container finished" podID="69fa6c84-911e-41f0-904d-1bd1b08d3b0c" containerID="fac65660cd60dff37d5a62bcebd54e995075ad4859a0003fc444546c0c091b9c" exitCode=0 Nov 24 14:01:07 crc kubenswrapper[4970]: I1124 14:01:07.860439 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"69fa6c84-911e-41f0-904d-1bd1b08d3b0c","Type":"ContainerDied","Data":"fac65660cd60dff37d5a62bcebd54e995075ad4859a0003fc444546c0c091b9c"} Nov 24 14:01:07 crc kubenswrapper[4970]: I1124 14:01:07.865624 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f0b6f151-1b32-4346-aabb-38ce2d89f50f","Type":"ContainerStarted","Data":"a49783d6db10bbc6c410db756b08158f0097715919d3c9fd1b5c7c1a59ba2feb"} Nov 24 14:01:07 crc kubenswrapper[4970]: I1124 14:01:07.922404 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.469574687 podStartE2EDuration="25.92238577s" podCreationTimestamp="2025-11-24 14:00:42 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.420278493 +0000 UTC m=+869.708035786" lastFinishedPulling="2025-11-24 14:01:01.873089576 +0000 UTC m=+877.160846869" observedRunningTime="2025-11-24 14:01:07.918942399 +0000 UTC m=+883.206699722" watchObservedRunningTime="2025-11-24 14:01:07.92238577 +0000 UTC m=+883.210143063" Nov 24 14:01:08 crc kubenswrapper[4970]: I1124 14:01:08.876129 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"69fa6c84-911e-41f0-904d-1bd1b08d3b0c","Type":"ContainerStarted","Data":"169a1103c55d26351abc14f4da2031e2d8e30bb19fc4616b1bb0b54d2a89d751"} Nov 24 14:01:08 crc kubenswrapper[4970]: I1124 14:01:08.901487 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.242530825 podStartE2EDuration="25.90146967s" podCreationTimestamp="2025-11-24 14:00:43 +0000 UTC" firstStartedPulling="2025-11-24 14:00:54.541349726 +0000 UTC m=+869.829107019" lastFinishedPulling="2025-11-24 14:01:01.200288571 +0000 UTC m=+876.488045864" observedRunningTime="2025-11-24 14:01:08.892933458 +0000 UTC m=+884.180690751" watchObservedRunningTime="2025-11-24 14:01:08.90146967 +0000 UTC m=+884.189226963" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.231992 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.232075 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.305231 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.532684 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.602362 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.884642 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.926409 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 14:01:09 crc kubenswrapper[4970]: I1124 14:01:09.950893 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.193376 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:10 crc kubenswrapper[4970]: E1124 14:01:10.193757 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="dnsmasq-dns" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.193782 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="dnsmasq-dns" Nov 24 14:01:10 crc kubenswrapper[4970]: E1124 14:01:10.193821 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="init" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.193832 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="init" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.193967 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="644eb975-00bf-42a7-ab6c-332544ea31d6" containerName="dnsmasq-dns" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.194771 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.197236 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.206534 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.222780 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.269821 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.269871 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.269904 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s92g2\" (UniqueName: \"kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.270031 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.371427 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.371499 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.371524 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.371562 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s92g2\" (UniqueName: \"kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.372405 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.372820 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.372968 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.398066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s92g2\" (UniqueName: \"kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2\") pod \"dnsmasq-dns-5848494dd9-c6n7f\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.400786 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-vz6qv"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.401997 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.404008 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.419350 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vz6qv"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.472962 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovs-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.473021 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1018f7-87e9-49d7-aa02-ee711b839d82-config\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.473046 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk6b8\" (UniqueName: \"kubernetes.io/projected/3f1018f7-87e9-49d7-aa02-ee711b839d82-kube-api-access-rk6b8\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.473073 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-combined-ca-bundle\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.473194 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovn-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.473216 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.511114 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.512406 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.516849 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.518662 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-lrxzn" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.519908 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.524386 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.525373 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.529022 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.545386 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577316 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovs-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577406 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-config\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577446 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1018f7-87e9-49d7-aa02-ee711b839d82-config\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577482 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk6b8\" (UniqueName: \"kubernetes.io/projected/3f1018f7-87e9-49d7-aa02-ee711b839d82-kube-api-access-rk6b8\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577549 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4995\" (UniqueName: \"kubernetes.io/projected/8561d773-9da8-4031-92c2-84ebd42a1afc-kube-api-access-z4995\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577599 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-combined-ca-bundle\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577630 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577655 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-scripts\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577687 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577787 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577831 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577867 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovn-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577685 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovs-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577902 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.577997 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3f1018f7-87e9-49d7-aa02-ee711b839d82-ovn-rundir\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.578155 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f1018f7-87e9-49d7-aa02-ee711b839d82-config\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.585662 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-combined-ca-bundle\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.590134 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f1018f7-87e9-49d7-aa02-ee711b839d82-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.593547 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.594849 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.596914 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk6b8\" (UniqueName: \"kubernetes.io/projected/3f1018f7-87e9-49d7-aa02-ee711b839d82-kube-api-access-rk6b8\") pod \"ovn-controller-metrics-vz6qv\" (UID: \"3f1018f7-87e9-49d7-aa02-ee711b839d82\") " pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.599707 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.609213 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679530 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4995\" (UniqueName: \"kubernetes.io/projected/8561d773-9da8-4031-92c2-84ebd42a1afc-kube-api-access-z4995\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679572 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679614 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-scripts\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679636 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679668 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkjn\" (UniqueName: \"kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679692 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679721 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679755 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679778 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679829 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679846 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-config\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.679867 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.680014 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.680907 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-scripts\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.681095 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8561d773-9da8-4031-92c2-84ebd42a1afc-config\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.684144 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.684268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.696238 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4995\" (UniqueName: \"kubernetes.io/projected/8561d773-9da8-4031-92c2-84ebd42a1afc-kube-api-access-z4995\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.697161 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8561d773-9da8-4031-92c2-84ebd42a1afc-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"8561d773-9da8-4031-92c2-84ebd42a1afc\") " pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.764659 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-vz6qv" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.780729 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkjn\" (UniqueName: \"kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.780775 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.780806 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.780884 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.780903 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.781820 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.781901 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.782024 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.782462 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.801846 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkjn\" (UniqueName: \"kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn\") pod \"dnsmasq-dns-5c7b6b5695-5m68w\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.839335 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 14:01:10 crc kubenswrapper[4970]: I1124 14:01:10.947595 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.033614 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.040063 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-vz6qv"] Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.203842 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.204841 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:01:11 crc kubenswrapper[4970]: W1124 14:01:11.368883 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8561d773_9da8_4031_92c2_84ebd42a1afc.slice/crio-163e3c206ec4ea332c8ad2e4b1f82c3fd9e259bbbc31b0e9c2e54f695e062cdb WatchSource:0}: Error finding container 163e3c206ec4ea332c8ad2e4b1f82c3fd9e259bbbc31b0e9c2e54f695e062cdb: Status 404 returned error can't find the container with id 163e3c206ec4ea332c8ad2e4b1f82c3fd9e259bbbc31b0e9c2e54f695e062cdb Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.370701 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.482985 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.901620 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vz6qv" event={"ID":"3f1018f7-87e9-49d7-aa02-ee711b839d82","Type":"ContainerStarted","Data":"a553a206aa17e9ae4a86d14472cf923941ef7522e942a602fb79d3990d94426b"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.901862 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-vz6qv" event={"ID":"3f1018f7-87e9-49d7-aa02-ee711b839d82","Type":"ContainerStarted","Data":"6a4f2e42d1e9593044ae4461b9c74a8cdbddd83f95a85ff6227d1e723dda9938"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.903302 4970 generic.go:334] "Generic (PLEG): container finished" podID="d4d65dda-0a3d-495d-b366-2dd063c14bef" containerID="e9ae91a716597c9b00d6dd2c2a88c508974a6d283d97aa83a3f3927792d66519" exitCode=0 Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.903416 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" event={"ID":"d4d65dda-0a3d-495d-b366-2dd063c14bef","Type":"ContainerDied","Data":"e9ae91a716597c9b00d6dd2c2a88c508974a6d283d97aa83a3f3927792d66519"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.903449 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" event={"ID":"d4d65dda-0a3d-495d-b366-2dd063c14bef","Type":"ContainerStarted","Data":"390d1a06e06a16c5ee693c6dd679403f67602df133c1ad44a6711255a8321b5d"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.916362 4970 generic.go:334] "Generic (PLEG): container finished" podID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerID="b015380d3d6b31b0b3559744ed78465d25afaa318c53196e2bcb243f2375b5c8" exitCode=0 Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.916480 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" event={"ID":"ac16f606-8254-4c99-bb2f-ecd992b10b8b","Type":"ContainerDied","Data":"b015380d3d6b31b0b3559744ed78465d25afaa318c53196e2bcb243f2375b5c8"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.916513 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" event={"ID":"ac16f606-8254-4c99-bb2f-ecd992b10b8b","Type":"ContainerStarted","Data":"3a1058b54b2101b8e020c19edd0cc66ccffab8693381dacdc0e0002fc47fcb10"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.926744 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8561d773-9da8-4031-92c2-84ebd42a1afc","Type":"ContainerStarted","Data":"163e3c206ec4ea332c8ad2e4b1f82c3fd9e259bbbc31b0e9c2e54f695e062cdb"} Nov 24 14:01:11 crc kubenswrapper[4970]: I1124 14:01:11.961035 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-vz6qv" podStartSLOduration=1.9610047640000001 podStartE2EDuration="1.961004764s" podCreationTimestamp="2025-11-24 14:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:11.922684327 +0000 UTC m=+887.210441620" watchObservedRunningTime="2025-11-24 14:01:11.961004764 +0000 UTC m=+887.248762067" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.226487 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.350169 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s92g2\" (UniqueName: \"kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2\") pod \"d4d65dda-0a3d-495d-b366-2dd063c14bef\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.350225 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config\") pod \"d4d65dda-0a3d-495d-b366-2dd063c14bef\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.350247 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb\") pod \"d4d65dda-0a3d-495d-b366-2dd063c14bef\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.350337 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc\") pod \"d4d65dda-0a3d-495d-b366-2dd063c14bef\" (UID: \"d4d65dda-0a3d-495d-b366-2dd063c14bef\") " Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.355113 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2" (OuterVolumeSpecName: "kube-api-access-s92g2") pod "d4d65dda-0a3d-495d-b366-2dd063c14bef" (UID: "d4d65dda-0a3d-495d-b366-2dd063c14bef"). InnerVolumeSpecName "kube-api-access-s92g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.370349 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config" (OuterVolumeSpecName: "config") pod "d4d65dda-0a3d-495d-b366-2dd063c14bef" (UID: "d4d65dda-0a3d-495d-b366-2dd063c14bef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.371311 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d4d65dda-0a3d-495d-b366-2dd063c14bef" (UID: "d4d65dda-0a3d-495d-b366-2dd063c14bef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.377280 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d4d65dda-0a3d-495d-b366-2dd063c14bef" (UID: "d4d65dda-0a3d-495d-b366-2dd063c14bef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.451964 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s92g2\" (UniqueName: \"kubernetes.io/projected/d4d65dda-0a3d-495d-b366-2dd063c14bef-kube-api-access-s92g2\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.452193 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.452268 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.452350 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d4d65dda-0a3d-495d-b366-2dd063c14bef-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.941401 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" event={"ID":"ac16f606-8254-4c99-bb2f-ecd992b10b8b","Type":"ContainerStarted","Data":"1a8de2725f8847e479321289e0332f040cf5d2ea5a761813034fe118214ba23d"} Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.943373 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.946736 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.948448 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5848494dd9-c6n7f" event={"ID":"d4d65dda-0a3d-495d-b366-2dd063c14bef","Type":"ContainerDied","Data":"390d1a06e06a16c5ee693c6dd679403f67602df133c1ad44a6711255a8321b5d"} Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.948651 4970 scope.go:117] "RemoveContainer" containerID="e9ae91a716597c9b00d6dd2c2a88c508974a6d283d97aa83a3f3927792d66519" Nov 24 14:01:12 crc kubenswrapper[4970]: I1124 14:01:12.980822 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" podStartSLOduration=2.980798534 podStartE2EDuration="2.980798534s" podCreationTimestamp="2025-11-24 14:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:12.968439303 +0000 UTC m=+888.256196626" watchObservedRunningTime="2025-11-24 14:01:12.980798534 +0000 UTC m=+888.268555847" Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.097999 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.104823 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5848494dd9-c6n7f"] Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.479758 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4d65dda-0a3d-495d-b366-2dd063c14bef" path="/var/lib/kubelet/pods/d4d65dda-0a3d-495d-b366-2dd063c14bef/volumes" Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.525071 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.525105 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 14:01:13 crc kubenswrapper[4970]: I1124 14:01:13.631796 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 14:01:14 crc kubenswrapper[4970]: I1124 14:01:14.076853 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 14:01:14 crc kubenswrapper[4970]: I1124 14:01:14.935816 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 14:01:14 crc kubenswrapper[4970]: I1124 14:01:14.936124 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 14:01:17 crc kubenswrapper[4970]: E1124 14:01:16.345410 4970 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:47688->38.102.83.12:33579: write tcp 38.102.83.12:47688->38.102.83.12:33579: write: broken pipe Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.838553 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.864052 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.864244 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="dnsmasq-dns" containerID="cri-o://1a8de2725f8847e479321289e0332f040cf5d2ea5a761813034fe118214ba23d" gracePeriod=10 Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.907407 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:01:17 crc kubenswrapper[4970]: E1124 14:01:16.907699 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4d65dda-0a3d-495d-b366-2dd063c14bef" containerName="init" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.907710 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4d65dda-0a3d-495d-b366-2dd063c14bef" containerName="init" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.907868 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4d65dda-0a3d-495d-b366-2dd063c14bef" containerName="init" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.908661 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:16.967012 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.034646 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.034694 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.034935 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkhpf\" (UniqueName: \"kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.034968 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.035000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.136475 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkhpf\" (UniqueName: \"kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.136569 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.136634 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.136724 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.136752 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.137500 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.137598 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.137665 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.137845 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.157354 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkhpf\" (UniqueName: \"kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf\") pod \"dnsmasq-dns-cf8bcbfcf-bwhb7\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.228316 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.901343 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:01:17 crc kubenswrapper[4970]: W1124 14:01:17.912359 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95ec9410_e3c8_4d9f_be06_6011987e3c07.slice/crio-d3a5a81ff75e28e9da579f81d5ba9b77a9b1c6c7053a0075f37f59e3050f1211 WatchSource:0}: Error finding container d3a5a81ff75e28e9da579f81d5ba9b77a9b1c6c7053a0075f37f59e3050f1211: Status 404 returned error can't find the container with id d3a5a81ff75e28e9da579f81d5ba9b77a9b1c6c7053a0075f37f59e3050f1211 Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.974698 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.980348 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.982426 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-z5f5w" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.982921 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.982948 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 24 14:01:17 crc kubenswrapper[4970]: I1124 14:01:17.983096 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.002911 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.005928 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerStarted","Data":"d3a5a81ff75e28e9da579f81d5ba9b77a9b1c6c7053a0075f37f59e3050f1211"} Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.026523 4970 generic.go:334] "Generic (PLEG): container finished" podID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerID="1a8de2725f8847e479321289e0332f040cf5d2ea5a761813034fe118214ba23d" exitCode=0 Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.026565 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" event={"ID":"ac16f606-8254-4c99-bb2f-ecd992b10b8b","Type":"ContainerDied","Data":"1a8de2725f8847e479321289e0332f040cf5d2ea5a761813034fe118214ba23d"} Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.052396 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-lock\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.052454 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jlnq\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-kube-api-access-6jlnq\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.052477 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.052505 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.052911 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-cache\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.154550 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.154774 4970 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.155793 4970 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.155853 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift podName:1fa4b14f-acb6-40eb-90ba-be3007901b3a nodeName:}" failed. No retries permitted until 2025-11-24 14:01:18.655833165 +0000 UTC m=+893.943590568 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift") pod "swift-storage-0" (UID: "1fa4b14f-acb6-40eb-90ba-be3007901b3a") : configmap "swift-ring-files" not found Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.155764 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.156004 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.156524 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-cache\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.156608 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-lock\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.156683 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jlnq\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-kube-api-access-6jlnq\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.157131 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-lock\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.157182 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/1fa4b14f-acb6-40eb-90ba-be3007901b3a-cache\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.172885 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jlnq\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-kube-api-access-6jlnq\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.176255 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.283412 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t72v9"] Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.284433 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.286480 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.287496 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.289967 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.297930 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t72v9"] Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360361 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360461 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmvp\" (UniqueName: \"kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360511 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360541 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360626 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360655 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.360706 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462124 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462235 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462286 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462352 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462441 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462515 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmvp\" (UniqueName: \"kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.462620 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.463200 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.464067 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.464510 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.469650 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.471419 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.471956 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.484921 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmvp\" (UniqueName: \"kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp\") pod \"swift-ring-rebalance-t72v9\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.606986 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:18 crc kubenswrapper[4970]: I1124 14:01:18.665844 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.666159 4970 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.666212 4970 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:01:18 crc kubenswrapper[4970]: E1124 14:01:18.666306 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift podName:1fa4b14f-acb6-40eb-90ba-be3007901b3a nodeName:}" failed. No retries permitted until 2025-11-24 14:01:19.666278015 +0000 UTC m=+894.954035348 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift") pod "swift-storage-0" (UID: "1fa4b14f-acb6-40eb-90ba-be3007901b3a") : configmap "swift-ring-files" not found Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.034910 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerStarted","Data":"d14f13fd9417d4d135eb70de7827fda658c8ddb1e927290180c252bf25cf6a91"} Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.690815 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:19 crc kubenswrapper[4970]: E1124 14:01:19.691024 4970 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:01:19 crc kubenswrapper[4970]: E1124 14:01:19.691058 4970 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:01:19 crc kubenswrapper[4970]: E1124 14:01:19.691118 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift podName:1fa4b14f-acb6-40eb-90ba-be3007901b3a nodeName:}" failed. No retries permitted until 2025-11-24 14:01:21.691099486 +0000 UTC m=+896.978856779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift") pod "swift-storage-0" (UID: "1fa4b14f-acb6-40eb-90ba-be3007901b3a") : configmap "swift-ring-files" not found Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.836125 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.905395 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config\") pod \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.905546 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb\") pod \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.905599 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc\") pod \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.905661 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjkjn\" (UniqueName: \"kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn\") pod \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.905693 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb\") pod \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\" (UID: \"ac16f606-8254-4c99-bb2f-ecd992b10b8b\") " Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.924545 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn" (OuterVolumeSpecName: "kube-api-access-rjkjn") pod "ac16f606-8254-4c99-bb2f-ecd992b10b8b" (UID: "ac16f606-8254-4c99-bb2f-ecd992b10b8b"). InnerVolumeSpecName "kube-api-access-rjkjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.971854 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac16f606-8254-4c99-bb2f-ecd992b10b8b" (UID: "ac16f606-8254-4c99-bb2f-ecd992b10b8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.985372 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac16f606-8254-4c99-bb2f-ecd992b10b8b" (UID: "ac16f606-8254-4c99-bb2f-ecd992b10b8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:19 crc kubenswrapper[4970]: I1124 14:01:19.987668 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac16f606-8254-4c99-bb2f-ecd992b10b8b" (UID: "ac16f606-8254-4c99-bb2f-ecd992b10b8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.008653 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.008685 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.008695 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjkjn\" (UniqueName: \"kubernetes.io/projected/ac16f606-8254-4c99-bb2f-ecd992b10b8b-kube-api-access-rjkjn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.008707 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.020791 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config" (OuterVolumeSpecName: "config") pod "ac16f606-8254-4c99-bb2f-ecd992b10b8b" (UID: "ac16f606-8254-4c99-bb2f-ecd992b10b8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.046322 4970 generic.go:334] "Generic (PLEG): container finished" podID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerID="d14f13fd9417d4d135eb70de7827fda658c8ddb1e927290180c252bf25cf6a91" exitCode=0 Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.046542 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerDied","Data":"d14f13fd9417d4d135eb70de7827fda658c8ddb1e927290180c252bf25cf6a91"} Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.050388 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8561d773-9da8-4031-92c2-84ebd42a1afc","Type":"ContainerStarted","Data":"9275e531cab2768574e134a6df931811731c9b0193239f8fc8f4083dc514bd0d"} Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.057629 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" event={"ID":"ac16f606-8254-4c99-bb2f-ecd992b10b8b","Type":"ContainerDied","Data":"3a1058b54b2101b8e020c19edd0cc66ccffab8693381dacdc0e0002fc47fcb10"} Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.057689 4970 scope.go:117] "RemoveContainer" containerID="1a8de2725f8847e479321289e0332f040cf5d2ea5a761813034fe118214ba23d" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.057752 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6b5695-5m68w" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.110417 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac16f606-8254-4c99-bb2f-ecd992b10b8b-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.154005 4970 scope.go:117] "RemoveContainer" containerID="b015380d3d6b31b0b3559744ed78465d25afaa318c53196e2bcb243f2375b5c8" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.200935 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.262283 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6b5695-5m68w"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.323256 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t72v9"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.396452 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.431701 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-3696-account-create-z754x"] Nov 24 14:01:20 crc kubenswrapper[4970]: E1124 14:01:20.432154 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="dnsmasq-dns" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.432171 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="dnsmasq-dns" Nov 24 14:01:20 crc kubenswrapper[4970]: E1124 14:01:20.432182 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="init" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.432189 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="init" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.432386 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" containerName="dnsmasq-dns" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.432998 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.436817 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.456836 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3696-account-create-z754x"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.480363 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cksvv"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.481587 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.486718 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cksvv"] Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.512686 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.537825 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvdkn\" (UniqueName: \"kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.537917 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.538088 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsjqm\" (UniqueName: \"kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.540042 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.641653 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.641716 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvdkn\" (UniqueName: \"kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.641850 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.641922 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsjqm\" (UniqueName: \"kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.642661 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.642670 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.660740 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvdkn\" (UniqueName: \"kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn\") pod \"glance-3696-account-create-z754x\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.662250 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsjqm\" (UniqueName: \"kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm\") pod \"glance-db-create-cksvv\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " pod="openstack/glance-db-create-cksvv" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.753558 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:20 crc kubenswrapper[4970]: I1124 14:01:20.800770 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cksvv" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.070118 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t72v9" event={"ID":"b596f962-fb31-4389-937e-f42750aeaf6a","Type":"ContainerStarted","Data":"9ff20483efbf8a0b93128f39ab23fb2978b98400a8f8f71f397239b462572050"} Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.076155 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerStarted","Data":"4d3f5f6844380e3314c4ad131865a2361f732989825564ea8b077095cfd2be4f"} Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.076213 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.078810 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"8561d773-9da8-4031-92c2-84ebd42a1afc","Type":"ContainerStarted","Data":"42b6ae87defb4a80bfdab6832bf6800e8f8cc433da0eb1f6c34fabce2755a917"} Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.078928 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.100803 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" podStartSLOduration=5.100766517 podStartE2EDuration="5.100766517s" podCreationTimestamp="2025-11-24 14:01:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:21.091244319 +0000 UTC m=+896.379001612" watchObservedRunningTime="2025-11-24 14:01:21.100766517 +0000 UTC m=+896.388523810" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.118166 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.731393721 podStartE2EDuration="11.118144369s" podCreationTimestamp="2025-11-24 14:01:10 +0000 UTC" firstStartedPulling="2025-11-24 14:01:11.371444175 +0000 UTC m=+886.659201478" lastFinishedPulling="2025-11-24 14:01:19.758194783 +0000 UTC m=+895.045952126" observedRunningTime="2025-11-24 14:01:21.111436264 +0000 UTC m=+896.399193557" watchObservedRunningTime="2025-11-24 14:01:21.118144369 +0000 UTC m=+896.405901672" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.265486 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-3696-account-create-z754x"] Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.339288 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cksvv"] Nov 24 14:01:21 crc kubenswrapper[4970]: W1124 14:01:21.346972 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcef88f7b_658e_4d42_b264_20532c99aa42.slice/crio-43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3 WatchSource:0}: Error finding container 43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3: Status 404 returned error can't find the container with id 43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3 Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.482151 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac16f606-8254-4c99-bb2f-ecd992b10b8b" path="/var/lib/kubelet/pods/ac16f606-8254-4c99-bb2f-ecd992b10b8b/volumes" Nov 24 14:01:21 crc kubenswrapper[4970]: I1124 14:01:21.762529 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:21 crc kubenswrapper[4970]: E1124 14:01:21.762780 4970 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:01:21 crc kubenswrapper[4970]: E1124 14:01:21.762824 4970 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:01:21 crc kubenswrapper[4970]: E1124 14:01:21.762887 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift podName:1fa4b14f-acb6-40eb-90ba-be3007901b3a nodeName:}" failed. No retries permitted until 2025-11-24 14:01:25.762866405 +0000 UTC m=+901.050623718 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift") pod "swift-storage-0" (UID: "1fa4b14f-acb6-40eb-90ba-be3007901b3a") : configmap "swift-ring-files" not found Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.085707 4970 generic.go:334] "Generic (PLEG): container finished" podID="6bbc61d9-f25b-47c6-879e-db93d0e543f2" containerID="ec359ee4bd5a0b507b92900e21efc01299987149a7222c5df7c2c8ff2a9d4a0d" exitCode=0 Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.085773 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3696-account-create-z754x" event={"ID":"6bbc61d9-f25b-47c6-879e-db93d0e543f2","Type":"ContainerDied","Data":"ec359ee4bd5a0b507b92900e21efc01299987149a7222c5df7c2c8ff2a9d4a0d"} Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.085798 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3696-account-create-z754x" event={"ID":"6bbc61d9-f25b-47c6-879e-db93d0e543f2","Type":"ContainerStarted","Data":"d4eff58f57214823163c27933b1e821b76206e20a2946780cffedb4954fbbcbb"} Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.089873 4970 generic.go:334] "Generic (PLEG): container finished" podID="cef88f7b-658e-4d42-b264-20532c99aa42" containerID="4a516ad9356643f201a8149cee73fe0c58824c178a6cc8930fe663d976a71627" exitCode=0 Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.089927 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cksvv" event={"ID":"cef88f7b-658e-4d42-b264-20532c99aa42","Type":"ContainerDied","Data":"4a516ad9356643f201a8149cee73fe0c58824c178a6cc8930fe663d976a71627"} Nov 24 14:01:22 crc kubenswrapper[4970]: I1124 14:01:22.089948 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cksvv" event={"ID":"cef88f7b-658e-4d42-b264-20532c99aa42","Type":"ContainerStarted","Data":"43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3"} Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.625323 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cksvv" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.633570 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.693936 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts\") pod \"cef88f7b-658e-4d42-b264-20532c99aa42\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.693995 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts\") pod \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.694032 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvdkn\" (UniqueName: \"kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn\") pod \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\" (UID: \"6bbc61d9-f25b-47c6-879e-db93d0e543f2\") " Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.694086 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsjqm\" (UniqueName: \"kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm\") pod \"cef88f7b-658e-4d42-b264-20532c99aa42\" (UID: \"cef88f7b-658e-4d42-b264-20532c99aa42\") " Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.694657 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cef88f7b-658e-4d42-b264-20532c99aa42" (UID: "cef88f7b-658e-4d42-b264-20532c99aa42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.695269 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6bbc61d9-f25b-47c6-879e-db93d0e543f2" (UID: "6bbc61d9-f25b-47c6-879e-db93d0e543f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.698076 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm" (OuterVolumeSpecName: "kube-api-access-gsjqm") pod "cef88f7b-658e-4d42-b264-20532c99aa42" (UID: "cef88f7b-658e-4d42-b264-20532c99aa42"). InnerVolumeSpecName "kube-api-access-gsjqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.700827 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn" (OuterVolumeSpecName: "kube-api-access-vvdkn") pod "6bbc61d9-f25b-47c6-879e-db93d0e543f2" (UID: "6bbc61d9-f25b-47c6-879e-db93d0e543f2"). InnerVolumeSpecName "kube-api-access-vvdkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.796323 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cef88f7b-658e-4d42-b264-20532c99aa42-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.796360 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6bbc61d9-f25b-47c6-879e-db93d0e543f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.796380 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvdkn\" (UniqueName: \"kubernetes.io/projected/6bbc61d9-f25b-47c6-879e-db93d0e543f2-kube-api-access-vvdkn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:23 crc kubenswrapper[4970]: I1124 14:01:23.796393 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsjqm\" (UniqueName: \"kubernetes.io/projected/cef88f7b-658e-4d42-b264-20532c99aa42-kube-api-access-gsjqm\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.131005 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-3696-account-create-z754x" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.131008 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-3696-account-create-z754x" event={"ID":"6bbc61d9-f25b-47c6-879e-db93d0e543f2","Type":"ContainerDied","Data":"d4eff58f57214823163c27933b1e821b76206e20a2946780cffedb4954fbbcbb"} Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.131104 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4eff58f57214823163c27933b1e821b76206e20a2946780cffedb4954fbbcbb" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.132885 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t72v9" event={"ID":"b596f962-fb31-4389-937e-f42750aeaf6a","Type":"ContainerStarted","Data":"5c55f630e1272c9f58de56c3c95e90fc215f4bcb74dc941abf75e98ff52f9f84"} Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.135811 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cksvv" event={"ID":"cef88f7b-658e-4d42-b264-20532c99aa42","Type":"ContainerDied","Data":"43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3"} Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.135841 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43763af866849272880d78f4c5e319427cd0dc08ea04c60d61f7ffc43d48b9d3" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.135859 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cksvv" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.157270 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t72v9" podStartSLOduration=3.003949086 podStartE2EDuration="6.157244412s" podCreationTimestamp="2025-11-24 14:01:18 +0000 UTC" firstStartedPulling="2025-11-24 14:01:20.334753854 +0000 UTC m=+895.622511147" lastFinishedPulling="2025-11-24 14:01:23.48804918 +0000 UTC m=+898.775806473" observedRunningTime="2025-11-24 14:01:24.154532942 +0000 UTC m=+899.442290245" watchObservedRunningTime="2025-11-24 14:01:24.157244412 +0000 UTC m=+899.445001725" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.741716 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-ptwhn"] Nov 24 14:01:24 crc kubenswrapper[4970]: E1124 14:01:24.742385 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef88f7b-658e-4d42-b264-20532c99aa42" containerName="mariadb-database-create" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.742401 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef88f7b-658e-4d42-b264-20532c99aa42" containerName="mariadb-database-create" Nov 24 14:01:24 crc kubenswrapper[4970]: E1124 14:01:24.742416 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bbc61d9-f25b-47c6-879e-db93d0e543f2" containerName="mariadb-account-create" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.742424 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bbc61d9-f25b-47c6-879e-db93d0e543f2" containerName="mariadb-account-create" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.742643 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef88f7b-658e-4d42-b264-20532c99aa42" containerName="mariadb-database-create" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.742664 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bbc61d9-f25b-47c6-879e-db93d0e543f2" containerName="mariadb-account-create" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.743345 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.748195 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ptwhn"] Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.836707 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.836800 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4dkz\" (UniqueName: \"kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.854232 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-33f1-account-create-92lmd"] Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.855532 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.858685 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.871826 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-33f1-account-create-92lmd"] Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.937740 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmn7j\" (UniqueName: \"kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.937897 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.937952 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.938069 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4dkz\" (UniqueName: \"kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.938632 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:24 crc kubenswrapper[4970]: I1124 14:01:24.964465 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4dkz\" (UniqueName: \"kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz\") pod \"keystone-db-create-ptwhn\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.039531 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.039705 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmn7j\" (UniqueName: \"kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.040478 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.049338 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cmgxd"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.050678 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.066591 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmn7j\" (UniqueName: \"kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j\") pod \"keystone-33f1-account-create-92lmd\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.067762 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cmgxd"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.070317 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.141595 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.141725 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zd7n\" (UniqueName: \"kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.171021 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.187634 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9297-account-create-xkv78"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.188916 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.194667 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.197907 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9297-account-create-xkv78"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.243893 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.243991 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.244061 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2b5m\" (UniqueName: \"kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.244322 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zd7n\" (UniqueName: \"kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.247400 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.266356 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zd7n\" (UniqueName: \"kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n\") pod \"placement-db-create-cmgxd\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.345861 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.345939 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2b5m\" (UniqueName: \"kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.347272 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.360744 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2b5m\" (UniqueName: \"kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m\") pod \"placement-9297-account-create-xkv78\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.445563 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.533949 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-ptwhn"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.546160 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.648959 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-33f1-account-create-92lmd"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.685732 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-fpfvv"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.686758 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.691330 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjmj8" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.692947 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.695618 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fpfvv"] Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.709630 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.752311 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.752400 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkzrh\" (UniqueName: \"kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.752535 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.752567 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.854205 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.854560 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.854607 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.854647 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkzrh\" (UniqueName: \"kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: E1124 14:01:25.854912 4970 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 24 14:01:25 crc kubenswrapper[4970]: E1124 14:01:25.854932 4970 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 24 14:01:25 crc kubenswrapper[4970]: E1124 14:01:25.854979 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift podName:1fa4b14f-acb6-40eb-90ba-be3007901b3a nodeName:}" failed. No retries permitted until 2025-11-24 14:01:33.854964252 +0000 UTC m=+909.142721545 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift") pod "swift-storage-0" (UID: "1fa4b14f-acb6-40eb-90ba-be3007901b3a") : configmap "swift-ring-files" not found Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.855039 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.862300 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.862974 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.870089 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkzrh\" (UniqueName: \"kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.870386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle\") pod \"glance-db-sync-fpfvv\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:25 crc kubenswrapper[4970]: I1124 14:01:25.878150 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cmgxd"] Nov 24 14:01:25 crc kubenswrapper[4970]: W1124 14:01:25.881709 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6377db5_fe8e_404b_98f2_7ad39a001b80.slice/crio-ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148 WatchSource:0}: Error finding container ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148: Status 404 returned error can't find the container with id ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148 Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.028187 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.137477 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9297-account-create-xkv78"] Nov 24 14:01:26 crc kubenswrapper[4970]: W1124 14:01:26.150843 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod279dafe8_9eae_4665_a2f2_781688a11b9b.slice/crio-b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095 WatchSource:0}: Error finding container b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095: Status 404 returned error can't find the container with id b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095 Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.158340 4970 generic.go:334] "Generic (PLEG): container finished" podID="ee934db6-19ef-4f00-b20a-498b1c229b4b" containerID="f96b6ea125f374eb8b3d0a56057b4d010823242f039d476d69f650c04b3453d1" exitCode=0 Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.158435 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-33f1-account-create-92lmd" event={"ID":"ee934db6-19ef-4f00-b20a-498b1c229b4b","Type":"ContainerDied","Data":"f96b6ea125f374eb8b3d0a56057b4d010823242f039d476d69f650c04b3453d1"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.158460 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-33f1-account-create-92lmd" event={"ID":"ee934db6-19ef-4f00-b20a-498b1c229b4b","Type":"ContainerStarted","Data":"f10a8ef7c69b4364a9d4f9a646c1d37c9cd84e050552c3acb4cbfa59012c978d"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.161356 4970 generic.go:334] "Generic (PLEG): container finished" podID="d8d73b83-954e-4da1-ae25-f975d3687eeb" containerID="43ab091a8e02ef1ed64d265cc863ea547c2be3f2d27d487b4427947eeb33f2ad" exitCode=0 Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.161448 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ptwhn" event={"ID":"d8d73b83-954e-4da1-ae25-f975d3687eeb","Type":"ContainerDied","Data":"43ab091a8e02ef1ed64d265cc863ea547c2be3f2d27d487b4427947eeb33f2ad"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.161485 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ptwhn" event={"ID":"d8d73b83-954e-4da1-ae25-f975d3687eeb","Type":"ContainerStarted","Data":"dba968c2d3534ee5adaa3e5c60593e0ac03ca8394f40599f9f654bffac75a366"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.163745 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cmgxd" event={"ID":"f6377db5-fe8e-404b-98f2-7ad39a001b80","Type":"ContainerStarted","Data":"90ed9af6d1858625b5b6043f8203a0ca6ef5a44e746ecaa6bb884b98e65a5e48"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.163786 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cmgxd" event={"ID":"f6377db5-fe8e-404b-98f2-7ad39a001b80","Type":"ContainerStarted","Data":"ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148"} Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.191330 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.209616 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cmgxd" podStartSLOduration=1.209598315 podStartE2EDuration="1.209598315s" podCreationTimestamp="2025-11-24 14:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:26.206302509 +0000 UTC m=+901.494059802" watchObservedRunningTime="2025-11-24 14:01:26.209598315 +0000 UTC m=+901.497355618" Nov 24 14:01:26 crc kubenswrapper[4970]: I1124 14:01:26.562552 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-fpfvv"] Nov 24 14:01:26 crc kubenswrapper[4970]: W1124 14:01:26.563304 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9ec76e6_76f2_4454_803d_5b5cb58c37d9.slice/crio-b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760 WatchSource:0}: Error finding container b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760: Status 404 returned error can't find the container with id b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760 Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.174491 4970 generic.go:334] "Generic (PLEG): container finished" podID="f6377db5-fe8e-404b-98f2-7ad39a001b80" containerID="90ed9af6d1858625b5b6043f8203a0ca6ef5a44e746ecaa6bb884b98e65a5e48" exitCode=0 Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.174536 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cmgxd" event={"ID":"f6377db5-fe8e-404b-98f2-7ad39a001b80","Type":"ContainerDied","Data":"90ed9af6d1858625b5b6043f8203a0ca6ef5a44e746ecaa6bb884b98e65a5e48"} Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.178917 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fpfvv" event={"ID":"c9ec76e6-76f2-4454-803d-5b5cb58c37d9","Type":"ContainerStarted","Data":"b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760"} Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.181170 4970 generic.go:334] "Generic (PLEG): container finished" podID="279dafe8-9eae-4665-a2f2-781688a11b9b" containerID="ace05e2c1c0446e587ac307db01dc573ec554abcdd8b3aa28c25de520a608914" exitCode=0 Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.181269 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9297-account-create-xkv78" event={"ID":"279dafe8-9eae-4665-a2f2-781688a11b9b","Type":"ContainerDied","Data":"ace05e2c1c0446e587ac307db01dc573ec554abcdd8b3aa28c25de520a608914"} Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.181324 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9297-account-create-xkv78" event={"ID":"279dafe8-9eae-4665-a2f2-781688a11b9b","Type":"ContainerStarted","Data":"b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095"} Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.229792 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.288900 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.289109 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="dnsmasq-dns" containerID="cri-o://1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1" gracePeriod=10 Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.712745 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.719587 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.786317 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts\") pod \"d8d73b83-954e-4da1-ae25-f975d3687eeb\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.786461 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4dkz\" (UniqueName: \"kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz\") pod \"d8d73b83-954e-4da1-ae25-f975d3687eeb\" (UID: \"d8d73b83-954e-4da1-ae25-f975d3687eeb\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.786489 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts\") pod \"ee934db6-19ef-4f00-b20a-498b1c229b4b\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.786513 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmn7j\" (UniqueName: \"kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j\") pod \"ee934db6-19ef-4f00-b20a-498b1c229b4b\" (UID: \"ee934db6-19ef-4f00-b20a-498b1c229b4b\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.787121 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d8d73b83-954e-4da1-ae25-f975d3687eeb" (UID: "d8d73b83-954e-4da1-ae25-f975d3687eeb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.787798 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee934db6-19ef-4f00-b20a-498b1c229b4b" (UID: "ee934db6-19ef-4f00-b20a-498b1c229b4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.792432 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz" (OuterVolumeSpecName: "kube-api-access-n4dkz") pod "d8d73b83-954e-4da1-ae25-f975d3687eeb" (UID: "d8d73b83-954e-4da1-ae25-f975d3687eeb"). InnerVolumeSpecName "kube-api-access-n4dkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.819318 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j" (OuterVolumeSpecName: "kube-api-access-qmn7j") pod "ee934db6-19ef-4f00-b20a-498b1c229b4b" (UID: "ee934db6-19ef-4f00-b20a-498b1c229b4b"). InnerVolumeSpecName "kube-api-access-qmn7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.842823 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888188 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config\") pod \"fef86e1d-5b44-4213-9adc-2da9c188f733\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888242 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc\") pod \"fef86e1d-5b44-4213-9adc-2da9c188f733\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888289 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k997\" (UniqueName: \"kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997\") pod \"fef86e1d-5b44-4213-9adc-2da9c188f733\" (UID: \"fef86e1d-5b44-4213-9adc-2da9c188f733\") " Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888803 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4dkz\" (UniqueName: \"kubernetes.io/projected/d8d73b83-954e-4da1-ae25-f975d3687eeb-kube-api-access-n4dkz\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888824 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee934db6-19ef-4f00-b20a-498b1c229b4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888835 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmn7j\" (UniqueName: \"kubernetes.io/projected/ee934db6-19ef-4f00-b20a-498b1c229b4b-kube-api-access-qmn7j\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.888848 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d8d73b83-954e-4da1-ae25-f975d3687eeb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.891782 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997" (OuterVolumeSpecName: "kube-api-access-9k997") pod "fef86e1d-5b44-4213-9adc-2da9c188f733" (UID: "fef86e1d-5b44-4213-9adc-2da9c188f733"). InnerVolumeSpecName "kube-api-access-9k997". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.928040 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fef86e1d-5b44-4213-9adc-2da9c188f733" (UID: "fef86e1d-5b44-4213-9adc-2da9c188f733"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.937034 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config" (OuterVolumeSpecName: "config") pod "fef86e1d-5b44-4213-9adc-2da9c188f733" (UID: "fef86e1d-5b44-4213-9adc-2da9c188f733"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.989993 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.990022 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fef86e1d-5b44-4213-9adc-2da9c188f733-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:27 crc kubenswrapper[4970]: I1124 14:01:27.990032 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k997\" (UniqueName: \"kubernetes.io/projected/fef86e1d-5b44-4213-9adc-2da9c188f733-kube-api-access-9k997\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.190282 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-33f1-account-create-92lmd" event={"ID":"ee934db6-19ef-4f00-b20a-498b1c229b4b","Type":"ContainerDied","Data":"f10a8ef7c69b4364a9d4f9a646c1d37c9cd84e050552c3acb4cbfa59012c978d"} Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.190328 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-33f1-account-create-92lmd" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.190340 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f10a8ef7c69b4364a9d4f9a646c1d37c9cd84e050552c3acb4cbfa59012c978d" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.192717 4970 generic.go:334] "Generic (PLEG): container finished" podID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerID="1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1" exitCode=0 Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.192820 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.193248 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" event={"ID":"fef86e1d-5b44-4213-9adc-2da9c188f733","Type":"ContainerDied","Data":"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1"} Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.193281 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-cqdjs" event={"ID":"fef86e1d-5b44-4213-9adc-2da9c188f733","Type":"ContainerDied","Data":"d814d8f1c01ebc28332dc18274a2e3f7fea904ba655da467c416c9743ba3da55"} Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.193300 4970 scope.go:117] "RemoveContainer" containerID="1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.195828 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-ptwhn" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.206791 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-ptwhn" event={"ID":"d8d73b83-954e-4da1-ae25-f975d3687eeb","Type":"ContainerDied","Data":"dba968c2d3534ee5adaa3e5c60593e0ac03ca8394f40599f9f654bffac75a366"} Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.206860 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dba968c2d3534ee5adaa3e5c60593e0ac03ca8394f40599f9f654bffac75a366" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.227218 4970 scope.go:117] "RemoveContainer" containerID="1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.235221 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.242422 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-cqdjs"] Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.264940 4970 scope.go:117] "RemoveContainer" containerID="1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1" Nov 24 14:01:28 crc kubenswrapper[4970]: E1124 14:01:28.265717 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1\": container with ID starting with 1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1 not found: ID does not exist" containerID="1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.265746 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1"} err="failed to get container status \"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1\": rpc error: code = NotFound desc = could not find container \"1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1\": container with ID starting with 1bd093bf950d93eb1097a68828d354b4fdbe149e6b190365c058de92661731a1 not found: ID does not exist" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.265769 4970 scope.go:117] "RemoveContainer" containerID="1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37" Nov 24 14:01:28 crc kubenswrapper[4970]: E1124 14:01:28.266131 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37\": container with ID starting with 1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37 not found: ID does not exist" containerID="1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.266151 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37"} err="failed to get container status \"1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37\": rpc error: code = NotFound desc = could not find container \"1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37\": container with ID starting with 1477ffd76f13203e8e73077c129ea56041475a108a67c23dad75c545ad5efa37 not found: ID does not exist" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.501935 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.598396 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts\") pod \"f6377db5-fe8e-404b-98f2-7ad39a001b80\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.598449 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zd7n\" (UniqueName: \"kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n\") pod \"f6377db5-fe8e-404b-98f2-7ad39a001b80\" (UID: \"f6377db5-fe8e-404b-98f2-7ad39a001b80\") " Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.599327 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6377db5-fe8e-404b-98f2-7ad39a001b80" (UID: "f6377db5-fe8e-404b-98f2-7ad39a001b80"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.602955 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n" (OuterVolumeSpecName: "kube-api-access-8zd7n") pod "f6377db5-fe8e-404b-98f2-7ad39a001b80" (UID: "f6377db5-fe8e-404b-98f2-7ad39a001b80"). InnerVolumeSpecName "kube-api-access-8zd7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.659998 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.700415 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts\") pod \"279dafe8-9eae-4665-a2f2-781688a11b9b\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.700560 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2b5m\" (UniqueName: \"kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m\") pod \"279dafe8-9eae-4665-a2f2-781688a11b9b\" (UID: \"279dafe8-9eae-4665-a2f2-781688a11b9b\") " Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.701007 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6377db5-fe8e-404b-98f2-7ad39a001b80-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.701041 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zd7n\" (UniqueName: \"kubernetes.io/projected/f6377db5-fe8e-404b-98f2-7ad39a001b80-kube-api-access-8zd7n\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.702527 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "279dafe8-9eae-4665-a2f2-781688a11b9b" (UID: "279dafe8-9eae-4665-a2f2-781688a11b9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.703915 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m" (OuterVolumeSpecName: "kube-api-access-w2b5m") pod "279dafe8-9eae-4665-a2f2-781688a11b9b" (UID: "279dafe8-9eae-4665-a2f2-781688a11b9b"). InnerVolumeSpecName "kube-api-access-w2b5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.802331 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2b5m\" (UniqueName: \"kubernetes.io/projected/279dafe8-9eae-4665-a2f2-781688a11b9b-kube-api-access-w2b5m\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:28 crc kubenswrapper[4970]: I1124 14:01:28.802374 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/279dafe8-9eae-4665-a2f2-781688a11b9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.206026 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cmgxd" event={"ID":"f6377db5-fe8e-404b-98f2-7ad39a001b80","Type":"ContainerDied","Data":"ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148"} Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.206081 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce6393aea3379d164eb39b599f08e9b3a51a92b0bb35e2a5da961c2cdd679148" Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.206051 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cmgxd" Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.212621 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9297-account-create-xkv78" event={"ID":"279dafe8-9eae-4665-a2f2-781688a11b9b","Type":"ContainerDied","Data":"b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095"} Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.212660 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b40d5a3a286360d571399164099e39ef2957985c5f2d6205cf4248420ad3c095" Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.212709 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9297-account-create-xkv78" Nov 24 14:01:29 crc kubenswrapper[4970]: I1124 14:01:29.482938 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" path="/var/lib/kubelet/pods/fef86e1d-5b44-4213-9adc-2da9c188f733/volumes" Nov 24 14:01:30 crc kubenswrapper[4970]: I1124 14:01:30.929652 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 14:01:31 crc kubenswrapper[4970]: I1124 14:01:31.233524 4970 generic.go:334] "Generic (PLEG): container finished" podID="b596f962-fb31-4389-937e-f42750aeaf6a" containerID="5c55f630e1272c9f58de56c3c95e90fc215f4bcb74dc941abf75e98ff52f9f84" exitCode=0 Nov 24 14:01:31 crc kubenswrapper[4970]: I1124 14:01:31.233645 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t72v9" event={"ID":"b596f962-fb31-4389-937e-f42750aeaf6a","Type":"ContainerDied","Data":"5c55f630e1272c9f58de56c3c95e90fc215f4bcb74dc941abf75e98ff52f9f84"} Nov 24 14:01:33 crc kubenswrapper[4970]: I1124 14:01:33.889866 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:33 crc kubenswrapper[4970]: I1124 14:01:33.897529 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1fa4b14f-acb6-40eb-90ba-be3007901b3a-etc-swift\") pod \"swift-storage-0\" (UID: \"1fa4b14f-acb6-40eb-90ba-be3007901b3a\") " pod="openstack/swift-storage-0" Nov 24 14:01:33 crc kubenswrapper[4970]: I1124 14:01:33.987820 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.279562 4970 generic.go:334] "Generic (PLEG): container finished" podID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerID="0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974" exitCode=0 Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.279816 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerDied","Data":"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974"} Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.369185 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-78b67" podUID="bb1eaff5-0428-4ab0-ae3b-21f192c95279" containerName="ovn-controller" probeResult="failure" output=< Nov 24 14:01:36 crc kubenswrapper[4970]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 14:01:36 crc kubenswrapper[4970]: > Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.403901 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.416256 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-r89nk" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.623131 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-78b67-config-pvjrc"] Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.623464 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="dnsmasq-dns" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.623477 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="dnsmasq-dns" Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.623491 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d73b83-954e-4da1-ae25-f975d3687eeb" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.623499 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d73b83-954e-4da1-ae25-f975d3687eeb" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.624567 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279dafe8-9eae-4665-a2f2-781688a11b9b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624604 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="279dafe8-9eae-4665-a2f2-781688a11b9b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.624627 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="init" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624634 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="init" Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.624655 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6377db5-fe8e-404b-98f2-7ad39a001b80" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624662 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6377db5-fe8e-404b-98f2-7ad39a001b80" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: E1124 14:01:36.624682 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee934db6-19ef-4f00-b20a-498b1c229b4b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624689 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee934db6-19ef-4f00-b20a-498b1c229b4b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624894 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="279dafe8-9eae-4665-a2f2-781688a11b9b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624916 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef86e1d-5b44-4213-9adc-2da9c188f733" containerName="dnsmasq-dns" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624931 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee934db6-19ef-4f00-b20a-498b1c229b4b" containerName="mariadb-account-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624939 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6377db5-fe8e-404b-98f2-7ad39a001b80" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.624955 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d73b83-954e-4da1-ae25-f975d3687eeb" containerName="mariadb-database-create" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.627909 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.640940 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.654068 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67-config-pvjrc"] Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.775562 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.775921 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.775955 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.775993 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.776027 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9wbw\" (UniqueName: \"kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.776069 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.876954 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877008 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877034 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877063 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877090 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9wbw\" (UniqueName: \"kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877124 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877260 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877310 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877381 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.877841 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.879075 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.907326 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9wbw\" (UniqueName: \"kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw\") pod \"ovn-controller-78b67-config-pvjrc\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:36 crc kubenswrapper[4970]: I1124 14:01:36.963646 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:37 crc kubenswrapper[4970]: I1124 14:01:37.289300 4970 generic.go:334] "Generic (PLEG): container finished" podID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerID="f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe" exitCode=0 Nov 24 14:01:37 crc kubenswrapper[4970]: I1124 14:01:37.289365 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerDied","Data":"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe"} Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.111063 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301141 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301527 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301555 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301619 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301645 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301716 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdmvp\" (UniqueName: \"kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.301769 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf\") pod \"b596f962-fb31-4389-937e-f42750aeaf6a\" (UID: \"b596f962-fb31-4389-937e-f42750aeaf6a\") " Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.302313 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.302948 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.309133 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp" (OuterVolumeSpecName: "kube-api-access-vdmvp") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "kube-api-access-vdmvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.315515 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.330135 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerStarted","Data":"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58"} Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.330438 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.334891 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67-config-pvjrc"] Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.339945 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts" (OuterVolumeSpecName: "scripts") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.340087 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.344321 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b596f962-fb31-4389-937e-f42750aeaf6a" (UID: "b596f962-fb31-4389-937e-f42750aeaf6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.346764 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerStarted","Data":"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f"} Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.347226 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.354129 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t72v9" event={"ID":"b596f962-fb31-4389-937e-f42750aeaf6a","Type":"ContainerDied","Data":"9ff20483efbf8a0b93128f39ab23fb2978b98400a8f8f71f397239b462572050"} Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.354177 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ff20483efbf8a0b93128f39ab23fb2978b98400a8f8f71f397239b462572050" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.354254 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t72v9" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.382280 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.107242889 podStartE2EDuration="58.382261458s" podCreationTimestamp="2025-11-24 14:00:40 +0000 UTC" firstStartedPulling="2025-11-24 14:00:52.648754132 +0000 UTC m=+867.936511425" lastFinishedPulling="2025-11-24 14:01:00.923772701 +0000 UTC m=+876.211529994" observedRunningTime="2025-11-24 14:01:38.377816132 +0000 UTC m=+913.665573435" watchObservedRunningTime="2025-11-24 14:01:38.382261458 +0000 UTC m=+913.670018751" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.402979 4970 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403005 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403016 4970 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403025 4970 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b596f962-fb31-4389-937e-f42750aeaf6a-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403032 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b596f962-fb31-4389-937e-f42750aeaf6a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403041 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdmvp\" (UniqueName: \"kubernetes.io/projected/b596f962-fb31-4389-937e-f42750aeaf6a-kube-api-access-vdmvp\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403049 4970 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b596f962-fb31-4389-937e-f42750aeaf6a-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.403228 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.137754058 podStartE2EDuration="58.403208003s" podCreationTimestamp="2025-11-24 14:00:40 +0000 UTC" firstStartedPulling="2025-11-24 14:00:53.658419339 +0000 UTC m=+868.946176642" lastFinishedPulling="2025-11-24 14:01:00.923873294 +0000 UTC m=+876.211630587" observedRunningTime="2025-11-24 14:01:38.401033487 +0000 UTC m=+913.688790780" watchObservedRunningTime="2025-11-24 14:01:38.403208003 +0000 UTC m=+913.690965296" Nov 24 14:01:38 crc kubenswrapper[4970]: E1124 14:01:38.423709 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb596f962_fb31_4389_937e_f42750aeaf6a.slice/crio-9ff20483efbf8a0b93128f39ab23fb2978b98400a8f8f71f397239b462572050\": RecentStats: unable to find data in memory cache]" Nov 24 14:01:38 crc kubenswrapper[4970]: I1124 14:01:38.453436 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.373310 4970 generic.go:334] "Generic (PLEG): container finished" podID="4e7c304d-5d64-4349-a003-7452eacc6e55" containerID="023a6dcfc6a9b3fe97be4a430419e54f0851c25121d4d43603098a17c4d21c3b" exitCode=0 Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.373682 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-pvjrc" event={"ID":"4e7c304d-5d64-4349-a003-7452eacc6e55","Type":"ContainerDied","Data":"023a6dcfc6a9b3fe97be4a430419e54f0851c25121d4d43603098a17c4d21c3b"} Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.373777 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-pvjrc" event={"ID":"4e7c304d-5d64-4349-a003-7452eacc6e55","Type":"ContainerStarted","Data":"2b6d5c81b14b0bf256c1e6c2203566204e44dba0616cc2abe5c0b2aad11017d1"} Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.376222 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"ec162cd244d3f44f4df1b65d8ac6b4eaa3cf7f0a5f302317ffaf771d3f480376"} Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.390891 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fpfvv" event={"ID":"c9ec76e6-76f2-4454-803d-5b5cb58c37d9","Type":"ContainerStarted","Data":"0590b4bc1a00e2b45b16494ec48cc4f6d6cc43c062db9ba88c2a7a53bf105265"} Nov 24 14:01:39 crc kubenswrapper[4970]: I1124 14:01:39.410819 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-fpfvv" podStartSLOduration=2.989941446 podStartE2EDuration="14.410799246s" podCreationTimestamp="2025-11-24 14:01:25 +0000 UTC" firstStartedPulling="2025-11-24 14:01:26.565701536 +0000 UTC m=+901.853458829" lastFinishedPulling="2025-11-24 14:01:37.986559296 +0000 UTC m=+913.274316629" observedRunningTime="2025-11-24 14:01:39.407095149 +0000 UTC m=+914.694852452" watchObservedRunningTime="2025-11-24 14:01:39.410799246 +0000 UTC m=+914.698556529" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.413616 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"4483860f1726c61fea76ee79b781c3b08ca451851986ea082012a563c9332b82"} Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.413952 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"1cc59ee42c78d45a7728ef4f6456ac4f1fafbbe470b186ad598c0a83005be4a2"} Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.792490 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954135 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954270 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954319 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954432 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954461 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9wbw\" (UniqueName: \"kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954492 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn\") pod \"4e7c304d-5d64-4349-a003-7452eacc6e55\" (UID: \"4e7c304d-5d64-4349-a003-7452eacc6e55\") " Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.954962 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.955004 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.955655 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.955694 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run" (OuterVolumeSpecName: "var-run") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.956321 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts" (OuterVolumeSpecName: "scripts") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:40 crc kubenswrapper[4970]: I1124 14:01:40.961643 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw" (OuterVolumeSpecName: "kube-api-access-b9wbw") pod "4e7c304d-5d64-4349-a003-7452eacc6e55" (UID: "4e7c304d-5d64-4349-a003-7452eacc6e55"). InnerVolumeSpecName "kube-api-access-b9wbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.055865 4970 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.056306 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e7c304d-5d64-4349-a003-7452eacc6e55-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.056315 4970 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.056326 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9wbw\" (UniqueName: \"kubernetes.io/projected/4e7c304d-5d64-4349-a003-7452eacc6e55-kube-api-access-b9wbw\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.056336 4970 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.056344 4970 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4e7c304d-5d64-4349-a003-7452eacc6e55-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.204294 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.204378 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.366851 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-78b67" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.428940 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"5ba55acb7550ddfcdc2a5f203b8ba6540247a8b05a3496f8f3bc3cb9b0e321ba"} Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.428999 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"e018696e8a94ecd41692c2479d59558f371d1a6e688e9e568efa1380933aeb30"} Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.434457 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-pvjrc" event={"ID":"4e7c304d-5d64-4349-a003-7452eacc6e55","Type":"ContainerDied","Data":"2b6d5c81b14b0bf256c1e6c2203566204e44dba0616cc2abe5c0b2aad11017d1"} Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.434566 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b6d5c81b14b0bf256c1e6c2203566204e44dba0616cc2abe5c0b2aad11017d1" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.434475 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-pvjrc" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.897886 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-78b67-config-pvjrc"] Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.922844 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-78b67-config-pvjrc"] Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.982358 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-78b67-config-dd44n"] Nov 24 14:01:41 crc kubenswrapper[4970]: E1124 14:01:41.983115 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e7c304d-5d64-4349-a003-7452eacc6e55" containerName="ovn-config" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.983136 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e7c304d-5d64-4349-a003-7452eacc6e55" containerName="ovn-config" Nov 24 14:01:41 crc kubenswrapper[4970]: E1124 14:01:41.983163 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b596f962-fb31-4389-937e-f42750aeaf6a" containerName="swift-ring-rebalance" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.983173 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b596f962-fb31-4389-937e-f42750aeaf6a" containerName="swift-ring-rebalance" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.983470 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b596f962-fb31-4389-937e-f42750aeaf6a" containerName="swift-ring-rebalance" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.983497 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e7c304d-5d64-4349-a003-7452eacc6e55" containerName="ovn-config" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.984326 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.987946 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 14:01:41 crc kubenswrapper[4970]: I1124 14:01:41.991440 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67-config-dd44n"] Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079346 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079400 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079478 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079507 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2qxd\" (UniqueName: \"kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079659 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.079878 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.181784 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.181958 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.182042 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.182077 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.182139 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.182198 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2qxd\" (UniqueName: \"kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.183149 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.183263 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.183316 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.183339 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.185100 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.215520 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2qxd\" (UniqueName: \"kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd\") pod \"ovn-controller-78b67-config-dd44n\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.309500 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.450845 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"adb4f4d8d49219a5ac7ef1c81cf6dd4ca38e50584fb2dadea6ffed81a7800d8f"} Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.450882 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"499f850a59dc8a674a36e38ef3f9750f30fea0e0aa4f4ac807264f4bb7afdb86"} Nov 24 14:01:42 crc kubenswrapper[4970]: I1124 14:01:42.814941 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-78b67-config-dd44n"] Nov 24 14:01:42 crc kubenswrapper[4970]: W1124 14:01:42.820913 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf79ed3f_b356_40dd_8b6d_82c8a4ab816f.slice/crio-ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c WatchSource:0}: Error finding container ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c: Status 404 returned error can't find the container with id ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.461032 4970 generic.go:334] "Generic (PLEG): container finished" podID="cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" containerID="efee003b2d183170e97231f87e1b0975182e3d3f192ac172e0de50d21de2c70b" exitCode=0 Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.461103 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-dd44n" event={"ID":"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f","Type":"ContainerDied","Data":"efee003b2d183170e97231f87e1b0975182e3d3f192ac172e0de50d21de2c70b"} Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.461133 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-dd44n" event={"ID":"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f","Type":"ContainerStarted","Data":"ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c"} Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.493197 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e7c304d-5d64-4349-a003-7452eacc6e55" path="/var/lib/kubelet/pods/4e7c304d-5d64-4349-a003-7452eacc6e55/volumes" Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.494764 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"1dbcd03cccecb2166d4314f70f7c810ceec8227adc03baa7a7fe40ed131d00dd"} Nov 24 14:01:43 crc kubenswrapper[4970]: I1124 14:01:43.494891 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"3949d3dcf8864fb688157c0233f762616f17a7a125108365e10a8a645fa7df29"} Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.488345 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"141cfa9cd6e4e9caec4f258bcd2ea0dbbc8d84b75f9f64518e098f7b0355d615"} Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.488926 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"0ee9bce69def5238c353625733d8bf252ae0480c86888bc0e140a6fcbc80224b"} Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.488944 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"b21aa32b901ca3262e72bf39debe40e8d367fc99b31d03dadca821a08d147f62"} Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.811228 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922053 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922153 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922195 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922244 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2qxd\" (UniqueName: \"kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922282 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922326 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run\") pod \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\" (UID: \"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f\") " Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922533 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922673 4970 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.922962 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.923009 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run" (OuterVolumeSpecName: "var-run") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.923075 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.923841 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts" (OuterVolumeSpecName: "scripts") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:44 crc kubenswrapper[4970]: I1124 14:01:44.927288 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd" (OuterVolumeSpecName: "kube-api-access-d2qxd") pod "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" (UID: "cf79ed3f-b356-40dd-8b6d-82c8a4ab816f"). InnerVolumeSpecName "kube-api-access-d2qxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.024373 4970 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.024406 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.024416 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2qxd\" (UniqueName: \"kubernetes.io/projected/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-kube-api-access-d2qxd\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.024428 4970 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.024437 4970 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.503288 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"7d29769c68177a6ee3af04e14630272bfaae3995ce268b79f3bb796a481917e1"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.503522 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"eee1542323e6c4fd1c0bcc7e9a0a9246dcf636879fc25732c989eff600d22adc"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.503532 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"54abc762a9bdd4b65aa3e55d1255d359bbcd099699d2d79d7837e56a94c52c65"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.503541 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"1fa4b14f-acb6-40eb-90ba-be3007901b3a","Type":"ContainerStarted","Data":"437e5c28904441b126d4ab976e379403759884f816f156044f29b1648ced4ac7"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.506793 4970 generic.go:334] "Generic (PLEG): container finished" podID="c9ec76e6-76f2-4454-803d-5b5cb58c37d9" containerID="0590b4bc1a00e2b45b16494ec48cc4f6d6cc43c062db9ba88c2a7a53bf105265" exitCode=0 Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.506834 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fpfvv" event={"ID":"c9ec76e6-76f2-4454-803d-5b5cb58c37d9","Type":"ContainerDied","Data":"0590b4bc1a00e2b45b16494ec48cc4f6d6cc43c062db9ba88c2a7a53bf105265"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.510988 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-78b67-config-dd44n" event={"ID":"cf79ed3f-b356-40dd-8b6d-82c8a4ab816f","Type":"ContainerDied","Data":"ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c"} Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.511015 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca99a614387bc69c5c3d089a7cf1c4fe1e1956f2d674077f7fbdfaa280f6386c" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.511050 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-78b67-config-dd44n" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.629323 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=24.256514154 podStartE2EDuration="29.629298954s" podCreationTimestamp="2025-11-24 14:01:16 +0000 UTC" firstStartedPulling="2025-11-24 14:01:38.466194263 +0000 UTC m=+913.753951556" lastFinishedPulling="2025-11-24 14:01:43.838979063 +0000 UTC m=+919.126736356" observedRunningTime="2025-11-24 14:01:45.619395656 +0000 UTC m=+920.907152949" watchObservedRunningTime="2025-11-24 14:01:45.629298954 +0000 UTC m=+920.917056257" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.943621 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-78b67-config-dd44n"] Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.956547 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-78b67-config-dd44n"] Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.965860 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:45 crc kubenswrapper[4970]: E1124 14:01:45.966184 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" containerName="ovn-config" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.966199 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" containerName="ovn-config" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.966347 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" containerName="ovn-config" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.967174 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:45 crc kubenswrapper[4970]: I1124 14:01:45.970477 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.009827 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046057 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046165 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vmgw\" (UniqueName: \"kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046243 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046263 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046291 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.046310 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.147936 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.147978 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vmgw\" (UniqueName: \"kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148052 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148073 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148100 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148117 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148790 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148809 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.148976 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.149300 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.150182 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.168630 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vmgw\" (UniqueName: \"kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw\") pod \"dnsmasq-dns-86c887b9fc-cz82x\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.331777 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.806344 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:46 crc kubenswrapper[4970]: I1124 14:01:46.989759 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.064185 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkzrh\" (UniqueName: \"kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh\") pod \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.064337 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data\") pod \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.064400 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle\") pod \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.064477 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data\") pod \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\" (UID: \"c9ec76e6-76f2-4454-803d-5b5cb58c37d9\") " Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.069117 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c9ec76e6-76f2-4454-803d-5b5cb58c37d9" (UID: "c9ec76e6-76f2-4454-803d-5b5cb58c37d9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.073751 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh" (OuterVolumeSpecName: "kube-api-access-nkzrh") pod "c9ec76e6-76f2-4454-803d-5b5cb58c37d9" (UID: "c9ec76e6-76f2-4454-803d-5b5cb58c37d9"). InnerVolumeSpecName "kube-api-access-nkzrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.100707 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c9ec76e6-76f2-4454-803d-5b5cb58c37d9" (UID: "c9ec76e6-76f2-4454-803d-5b5cb58c37d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.117733 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data" (OuterVolumeSpecName: "config-data") pod "c9ec76e6-76f2-4454-803d-5b5cb58c37d9" (UID: "c9ec76e6-76f2-4454-803d-5b5cb58c37d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.166554 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.166964 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkzrh\" (UniqueName: \"kubernetes.io/projected/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-kube-api-access-nkzrh\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.166984 4970 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.167000 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9ec76e6-76f2-4454-803d-5b5cb58c37d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.489444 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf79ed3f-b356-40dd-8b6d-82c8a4ab816f" path="/var/lib/kubelet/pods/cf79ed3f-b356-40dd-8b6d-82c8a4ab816f/volumes" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.536119 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-fpfvv" event={"ID":"c9ec76e6-76f2-4454-803d-5b5cb58c37d9","Type":"ContainerDied","Data":"b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760"} Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.536253 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5aa1ea2c2260be915b81c08fbbced7145e79fa3d775bc7b6cd91377c8030760" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.536353 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-fpfvv" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.541884 4970 generic.go:334] "Generic (PLEG): container finished" podID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerID="c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9" exitCode=0 Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.541934 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" event={"ID":"567036ff-92e5-4cad-a405-86f39b63a2c5","Type":"ContainerDied","Data":"c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9"} Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.542040 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" event={"ID":"567036ff-92e5-4cad-a405-86f39b63a2c5","Type":"ContainerStarted","Data":"7e583db6d451e8c365701aad8f4ef0f77e6aca3ca38ec9b8f779ffda50df1785"} Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.951049 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.973792 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:01:47 crc kubenswrapper[4970]: E1124 14:01:47.974103 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ec76e6-76f2-4454-803d-5b5cb58c37d9" containerName="glance-db-sync" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.974121 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ec76e6-76f2-4454-803d-5b5cb58c37d9" containerName="glance-db-sync" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.974324 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ec76e6-76f2-4454-803d-5b5cb58c37d9" containerName="glance-db-sync" Nov 24 14:01:47 crc kubenswrapper[4970]: I1124 14:01:47.975177 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.017488 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089237 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089301 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089334 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089465 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089615 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv826\" (UniqueName: \"kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.089643 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190623 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190698 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190739 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190771 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190811 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv826\" (UniqueName: \"kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.190832 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.192140 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.192137 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.192299 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.192374 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.192541 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.212715 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv826\" (UniqueName: \"kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826\") pod \"dnsmasq-dns-cb6b7c77c-5jz68\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.295987 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.566379 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" event={"ID":"567036ff-92e5-4cad-a405-86f39b63a2c5","Type":"ContainerStarted","Data":"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d"} Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.569109 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.726171 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" podStartSLOduration=3.72614746 podStartE2EDuration="3.72614746s" podCreationTimestamp="2025-11-24 14:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:48.592395258 +0000 UTC m=+923.880152591" watchObservedRunningTime="2025-11-24 14:01:48.72614746 +0000 UTC m=+924.013904753" Nov 24 14:01:48 crc kubenswrapper[4970]: I1124 14:01:48.732988 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:01:49 crc kubenswrapper[4970]: I1124 14:01:49.575812 4970 generic.go:334] "Generic (PLEG): container finished" podID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerID="8ec867edc0e42e24e0847720796430851695c4457627c6cbb2755a34da91f59a" exitCode=0 Nov 24 14:01:49 crc kubenswrapper[4970]: I1124 14:01:49.575915 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" event={"ID":"e2bba80e-8ca2-4b12-b404-cac67c1a69e6","Type":"ContainerDied","Data":"8ec867edc0e42e24e0847720796430851695c4457627c6cbb2755a34da91f59a"} Nov 24 14:01:49 crc kubenswrapper[4970]: I1124 14:01:49.576180 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" event={"ID":"e2bba80e-8ca2-4b12-b404-cac67c1a69e6","Type":"ContainerStarted","Data":"2ee88eef8fcfc1a9d7d4c73e31ecdaa0c4fd7f9558729cdc5bc5caccf8697e0c"} Nov 24 14:01:49 crc kubenswrapper[4970]: I1124 14:01:49.576330 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="dnsmasq-dns" containerID="cri-o://06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d" gracePeriod=10 Nov 24 14:01:49 crc kubenswrapper[4970]: I1124 14:01:49.933224 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.032918 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.033344 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.033395 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.033448 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vmgw\" (UniqueName: \"kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.033473 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.033492 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc\") pod \"567036ff-92e5-4cad-a405-86f39b63a2c5\" (UID: \"567036ff-92e5-4cad-a405-86f39b63a2c5\") " Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.038343 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw" (OuterVolumeSpecName: "kube-api-access-2vmgw") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "kube-api-access-2vmgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.073383 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config" (OuterVolumeSpecName: "config") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.075163 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.075825 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.079479 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.080427 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "567036ff-92e5-4cad-a405-86f39b63a2c5" (UID: "567036ff-92e5-4cad-a405-86f39b63a2c5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135114 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135155 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135168 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135178 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vmgw\" (UniqueName: \"kubernetes.io/projected/567036ff-92e5-4cad-a405-86f39b63a2c5-kube-api-access-2vmgw\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135194 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.135205 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/567036ff-92e5-4cad-a405-86f39b63a2c5-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.585999 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" event={"ID":"e2bba80e-8ca2-4b12-b404-cac67c1a69e6","Type":"ContainerStarted","Data":"1308215d54a7e659b575a86127109b865bd51732edbe2ae8c3e9a5534ad45258"} Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.586235 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.590525 4970 generic.go:334] "Generic (PLEG): container finished" podID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerID="06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d" exitCode=0 Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.590566 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.590599 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" event={"ID":"567036ff-92e5-4cad-a405-86f39b63a2c5","Type":"ContainerDied","Data":"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d"} Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.590725 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86c887b9fc-cz82x" event={"ID":"567036ff-92e5-4cad-a405-86f39b63a2c5","Type":"ContainerDied","Data":"7e583db6d451e8c365701aad8f4ef0f77e6aca3ca38ec9b8f779ffda50df1785"} Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.590751 4970 scope.go:117] "RemoveContainer" containerID="06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.619717 4970 scope.go:117] "RemoveContainer" containerID="c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.625318 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" podStartSLOduration=3.625298175 podStartE2EDuration="3.625298175s" podCreationTimestamp="2025-11-24 14:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:50.620083339 +0000 UTC m=+925.907840642" watchObservedRunningTime="2025-11-24 14:01:50.625298175 +0000 UTC m=+925.913055468" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.643711 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.644958 4970 scope.go:117] "RemoveContainer" containerID="06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d" Nov 24 14:01:50 crc kubenswrapper[4970]: E1124 14:01:50.645821 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d\": container with ID starting with 06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d not found: ID does not exist" containerID="06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.645904 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d"} err="failed to get container status \"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d\": rpc error: code = NotFound desc = could not find container \"06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d\": container with ID starting with 06df713942fb4378c7a3f4a23a22506f44478119e5bf880d9fe31fcaffaa728d not found: ID does not exist" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.645949 4970 scope.go:117] "RemoveContainer" containerID="c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9" Nov 24 14:01:50 crc kubenswrapper[4970]: E1124 14:01:50.646297 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9\": container with ID starting with c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9 not found: ID does not exist" containerID="c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.646349 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9"} err="failed to get container status \"c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9\": rpc error: code = NotFound desc = could not find container \"c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9\": container with ID starting with c15bab90340f05dd05179ec8e066a8fe2f00bbd9803c287f61a6c05b3a0c1dd9 not found: ID does not exist" Nov 24 14:01:50 crc kubenswrapper[4970]: I1124 14:01:50.652179 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86c887b9fc-cz82x"] Nov 24 14:01:51 crc kubenswrapper[4970]: I1124 14:01:51.484880 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" path="/var/lib/kubelet/pods/567036ff-92e5-4cad-a405-86f39b63a2c5/volumes" Nov 24 14:01:51 crc kubenswrapper[4970]: I1124 14:01:51.931846 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.253810 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.280797 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-z5stn"] Nov 24 14:01:52 crc kubenswrapper[4970]: E1124 14:01:52.281286 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="init" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.281311 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="init" Nov 24 14:01:52 crc kubenswrapper[4970]: E1124 14:01:52.281343 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="dnsmasq-dns" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.281351 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="dnsmasq-dns" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.281605 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="567036ff-92e5-4cad-a405-86f39b63a2c5" containerName="dnsmasq-dns" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.282537 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.291140 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z5stn"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.373647 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-k696g"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.374752 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.383759 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll7l8\" (UniqueName: \"kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.383837 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.393057 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5205-account-create-2qjmw"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.397029 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.403053 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.415228 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5205-account-create-2qjmw"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.436608 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-k696g"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.478623 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6189-account-create-5fm2f"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.479996 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.481845 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485697 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvkp\" (UniqueName: \"kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485758 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485837 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485872 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6grg\" (UniqueName: \"kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485902 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll7l8\" (UniqueName: \"kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.485979 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.486455 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.501010 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6189-account-create-5fm2f"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.510818 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll7l8\" (UniqueName: \"kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8\") pod \"cinder-db-create-z5stn\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.574858 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-wkwfz"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.575884 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587270 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587318 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587342 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvkp\" (UniqueName: \"kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587370 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9cg6\" (UniqueName: \"kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587468 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.587496 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6grg\" (UniqueName: \"kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.588244 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wkwfz"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.588520 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.589302 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.601782 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.614158 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6grg\" (UniqueName: \"kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg\") pod \"barbican-5205-account-create-2qjmw\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.615316 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvkp\" (UniqueName: \"kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp\") pod \"barbican-db-create-k696g\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.651359 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c6tvk"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.652346 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.656750 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ptrd5" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.657077 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.657093 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.658799 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.666512 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c6tvk"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689454 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r685f\" (UniqueName: \"kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689545 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689605 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689632 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9cg6\" (UniqueName: \"kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689673 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qwp2\" (UniqueName: \"kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689703 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.689742 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.690639 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.703186 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-k696g" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.707701 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8446-account-create-kqdzx"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.708652 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.712399 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.715025 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9cg6\" (UniqueName: \"kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6\") pod \"cinder-6189-account-create-5fm2f\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.719945 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.727402 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8446-account-create-kqdzx"] Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791683 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791767 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qwp2\" (UniqueName: \"kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791823 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791850 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791939 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.791991 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r685f\" (UniqueName: \"kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.792014 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmm5h\" (UniqueName: \"kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.792784 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.793226 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.795201 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.796930 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.815844 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r685f\" (UniqueName: \"kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f\") pod \"keystone-db-sync-c6tvk\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.831125 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qwp2\" (UniqueName: \"kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2\") pod \"neutron-db-create-wkwfz\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.892183 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.893153 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.893427 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmm5h\" (UniqueName: \"kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.894253 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:52 crc kubenswrapper[4970]: I1124 14:01:52.910091 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmm5h\" (UniqueName: \"kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h\") pod \"neutron-8446-account-create-kqdzx\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.012195 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.117801 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.161588 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-k696g"] Nov 24 14:01:53 crc kubenswrapper[4970]: W1124 14:01:53.169281 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cf9ba03_18af_4c11_ad00_d6e26ef65c8f.slice/crio-ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7 WatchSource:0}: Error finding container ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7: Status 404 returned error can't find the container with id ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7 Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.247087 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-z5stn"] Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.306613 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5205-account-create-2qjmw"] Nov 24 14:01:53 crc kubenswrapper[4970]: W1124 14:01:53.320681 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a055662_deb7_4f6a_bef1_a69e71f155e0.slice/crio-622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1 WatchSource:0}: Error finding container 622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1: Status 404 returned error can't find the container with id 622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1 Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.399021 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6189-account-create-5fm2f"] Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.486742 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-wkwfz"] Nov 24 14:01:53 crc kubenswrapper[4970]: W1124 14:01:53.490962 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d1ae5b1_853f_4c6c_9a9f_d2e6db92a489.slice/crio-70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9 WatchSource:0}: Error finding container 70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9: Status 404 returned error can't find the container with id 70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9 Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.587950 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c6tvk"] Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.633423 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wkwfz" event={"ID":"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489","Type":"ContainerStarted","Data":"70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.635480 4970 generic.go:334] "Generic (PLEG): container finished" podID="9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" containerID="31b20bb8ad269e9ebd09a8f86d9da47dd7aa7f4dad9109e7e2e63c54f61df4f6" exitCode=0 Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.635523 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-k696g" event={"ID":"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f","Type":"ContainerDied","Data":"31b20bb8ad269e9ebd09a8f86d9da47dd7aa7f4dad9109e7e2e63c54f61df4f6"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.635540 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-k696g" event={"ID":"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f","Type":"ContainerStarted","Data":"ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.639711 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6tvk" event={"ID":"145d1a69-5b39-4ec0-8544-3034bded1955","Type":"ContainerStarted","Data":"b392c5b213095248b217435c5067684d1a3919b44ca5beba0e6349b924f6eded"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.652531 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5205-account-create-2qjmw" event={"ID":"8a055662-deb7-4f6a-bef1-a69e71f155e0","Type":"ContainerStarted","Data":"dcce86608a5d568912915abfef47037ffee042a6dccceb3621416d2e2deafd02"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.652602 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5205-account-create-2qjmw" event={"ID":"8a055662-deb7-4f6a-bef1-a69e71f155e0","Type":"ContainerStarted","Data":"622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.663090 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z5stn" event={"ID":"d77cc176-a7dd-4510-8f10-a5cc227ca06f","Type":"ContainerStarted","Data":"33c6643207a43500f6e9e52f3b50294535ed8997fc0dc80ceb6bc37b295d350a"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.663149 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z5stn" event={"ID":"d77cc176-a7dd-4510-8f10-a5cc227ca06f","Type":"ContainerStarted","Data":"c3251b191bec65f3d53d8ab08196b981049c2d47881cc49f02915f25d7408bc1"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.674944 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8446-account-create-kqdzx"] Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.675767 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6189-account-create-5fm2f" event={"ID":"c94b8e19-9edb-4a82-95a9-5717afa5ae6d","Type":"ContainerStarted","Data":"831d65e5cc0dae3cdedb599d67ee83ce3a03c00b2e6a1ee5a8408fb97706a679"} Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.677170 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-5205-account-create-2qjmw" podStartSLOduration=1.677160169 podStartE2EDuration="1.677160169s" podCreationTimestamp="2025-11-24 14:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:53.676214415 +0000 UTC m=+928.963971718" watchObservedRunningTime="2025-11-24 14:01:53.677160169 +0000 UTC m=+928.964917462" Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.702560 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-6189-account-create-5fm2f" podStartSLOduration=1.70254541 podStartE2EDuration="1.70254541s" podCreationTimestamp="2025-11-24 14:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:53.701786801 +0000 UTC m=+928.989544104" watchObservedRunningTime="2025-11-24 14:01:53.70254541 +0000 UTC m=+928.990302693" Nov 24 14:01:53 crc kubenswrapper[4970]: W1124 14:01:53.704766 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd54ba95_cb56_4068_90c9_06cfd72111c4.slice/crio-ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2 WatchSource:0}: Error finding container ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2: Status 404 returned error can't find the container with id ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2 Nov 24 14:01:53 crc kubenswrapper[4970]: I1124 14:01:53.727882 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-z5stn" podStartSLOduration=1.727858889 podStartE2EDuration="1.727858889s" podCreationTimestamp="2025-11-24 14:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:01:53.715495618 +0000 UTC m=+929.003252901" watchObservedRunningTime="2025-11-24 14:01:53.727858889 +0000 UTC m=+929.015616182" Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.686024 4970 generic.go:334] "Generic (PLEG): container finished" podID="dd54ba95-cb56-4068-90c9-06cfd72111c4" containerID="67129bc7ccb47df1a286fb0ff31884279c3c06c3697633e07d8bc857a1b389ba" exitCode=0 Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.686370 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8446-account-create-kqdzx" event={"ID":"dd54ba95-cb56-4068-90c9-06cfd72111c4","Type":"ContainerDied","Data":"67129bc7ccb47df1a286fb0ff31884279c3c06c3697633e07d8bc857a1b389ba"} Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.686396 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8446-account-create-kqdzx" event={"ID":"dd54ba95-cb56-4068-90c9-06cfd72111c4","Type":"ContainerStarted","Data":"ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2"} Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.688301 4970 generic.go:334] "Generic (PLEG): container finished" podID="c94b8e19-9edb-4a82-95a9-5717afa5ae6d" containerID="994745c33b0ef29b802f527e53366b3bdb6621cc0946896116b5d409c5119325" exitCode=0 Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.688352 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6189-account-create-5fm2f" event={"ID":"c94b8e19-9edb-4a82-95a9-5717afa5ae6d","Type":"ContainerDied","Data":"994745c33b0ef29b802f527e53366b3bdb6621cc0946896116b5d409c5119325"} Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.690549 4970 generic.go:334] "Generic (PLEG): container finished" podID="7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" containerID="9f8f5954bc3a8e93e9bbb1263f0b358632b019b48093d84865515af5bd21dbbb" exitCode=0 Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.690587 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wkwfz" event={"ID":"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489","Type":"ContainerDied","Data":"9f8f5954bc3a8e93e9bbb1263f0b358632b019b48093d84865515af5bd21dbbb"} Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.692077 4970 generic.go:334] "Generic (PLEG): container finished" podID="8a055662-deb7-4f6a-bef1-a69e71f155e0" containerID="dcce86608a5d568912915abfef47037ffee042a6dccceb3621416d2e2deafd02" exitCode=0 Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.692141 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5205-account-create-2qjmw" event={"ID":"8a055662-deb7-4f6a-bef1-a69e71f155e0","Type":"ContainerDied","Data":"dcce86608a5d568912915abfef47037ffee042a6dccceb3621416d2e2deafd02"} Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.693761 4970 generic.go:334] "Generic (PLEG): container finished" podID="d77cc176-a7dd-4510-8f10-a5cc227ca06f" containerID="33c6643207a43500f6e9e52f3b50294535ed8997fc0dc80ceb6bc37b295d350a" exitCode=0 Nov 24 14:01:54 crc kubenswrapper[4970]: I1124 14:01:54.693884 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z5stn" event={"ID":"d77cc176-a7dd-4510-8f10-a5cc227ca06f","Type":"ContainerDied","Data":"33c6643207a43500f6e9e52f3b50294535ed8997fc0dc80ceb6bc37b295d350a"} Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.041094 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-k696g" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.133887 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvkp\" (UniqueName: \"kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp\") pod \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.133958 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts\") pod \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\" (UID: \"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f\") " Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.134950 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" (UID: "9cf9ba03-18af-4c11-ad00-d6e26ef65c8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.140060 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp" (OuterVolumeSpecName: "kube-api-access-xhvkp") pod "9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" (UID: "9cf9ba03-18af-4c11-ad00-d6e26ef65c8f"). InnerVolumeSpecName "kube-api-access-xhvkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.235467 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvkp\" (UniqueName: \"kubernetes.io/projected/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-kube-api-access-xhvkp\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.235498 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.704099 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-k696g" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.704284 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-k696g" event={"ID":"9cf9ba03-18af-4c11-ad00-d6e26ef65c8f","Type":"ContainerDied","Data":"ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7"} Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.704386 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce4b77a698c9db6f2d2cf7d2e93ecbc13d2c865d238bf2087ef94575f9b8f9d7" Nov 24 14:01:55 crc kubenswrapper[4970]: I1124 14:01:55.992214 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.047674 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmm5h\" (UniqueName: \"kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h\") pod \"dd54ba95-cb56-4068-90c9-06cfd72111c4\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.048072 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts\") pod \"dd54ba95-cb56-4068-90c9-06cfd72111c4\" (UID: \"dd54ba95-cb56-4068-90c9-06cfd72111c4\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.049165 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd54ba95-cb56-4068-90c9-06cfd72111c4" (UID: "dd54ba95-cb56-4068-90c9-06cfd72111c4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.057321 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h" (OuterVolumeSpecName: "kube-api-access-cmm5h") pod "dd54ba95-cb56-4068-90c9-06cfd72111c4" (UID: "dd54ba95-cb56-4068-90c9-06cfd72111c4"). InnerVolumeSpecName "kube-api-access-cmm5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.150145 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmm5h\" (UniqueName: \"kubernetes.io/projected/dd54ba95-cb56-4068-90c9-06cfd72111c4-kube-api-access-cmm5h\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.150182 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd54ba95-cb56-4068-90c9-06cfd72111c4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.261757 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.277175 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.281804 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.291203 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.352796 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts\") pod \"8a055662-deb7-4f6a-bef1-a69e71f155e0\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353117 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6grg\" (UniqueName: \"kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg\") pod \"8a055662-deb7-4f6a-bef1-a69e71f155e0\" (UID: \"8a055662-deb7-4f6a-bef1-a69e71f155e0\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353154 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts\") pod \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353189 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9cg6\" (UniqueName: \"kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6\") pod \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353257 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qwp2\" (UniqueName: \"kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2\") pod \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\" (UID: \"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353305 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll7l8\" (UniqueName: \"kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8\") pod \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353326 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts\") pod \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\" (UID: \"d77cc176-a7dd-4510-8f10-a5cc227ca06f\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.353383 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts\") pod \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\" (UID: \"c94b8e19-9edb-4a82-95a9-5717afa5ae6d\") " Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.354117 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c94b8e19-9edb-4a82-95a9-5717afa5ae6d" (UID: "c94b8e19-9edb-4a82-95a9-5717afa5ae6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.354468 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a055662-deb7-4f6a-bef1-a69e71f155e0" (UID: "8a055662-deb7-4f6a-bef1-a69e71f155e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.359250 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" (UID: "7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.361106 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2" (OuterVolumeSpecName: "kube-api-access-6qwp2") pod "7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" (UID: "7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489"). InnerVolumeSpecName "kube-api-access-6qwp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.365879 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6" (OuterVolumeSpecName: "kube-api-access-b9cg6") pod "c94b8e19-9edb-4a82-95a9-5717afa5ae6d" (UID: "c94b8e19-9edb-4a82-95a9-5717afa5ae6d"). InnerVolumeSpecName "kube-api-access-b9cg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.366020 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8" (OuterVolumeSpecName: "kube-api-access-ll7l8") pod "d77cc176-a7dd-4510-8f10-a5cc227ca06f" (UID: "d77cc176-a7dd-4510-8f10-a5cc227ca06f"). InnerVolumeSpecName "kube-api-access-ll7l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.366154 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d77cc176-a7dd-4510-8f10-a5cc227ca06f" (UID: "d77cc176-a7dd-4510-8f10-a5cc227ca06f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.366328 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg" (OuterVolumeSpecName: "kube-api-access-k6grg") pod "8a055662-deb7-4f6a-bef1-a69e71f155e0" (UID: "8a055662-deb7-4f6a-bef1-a69e71f155e0"). InnerVolumeSpecName "kube-api-access-k6grg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454790 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454823 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a055662-deb7-4f6a-bef1-a69e71f155e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454832 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6grg\" (UniqueName: \"kubernetes.io/projected/8a055662-deb7-4f6a-bef1-a69e71f155e0-kube-api-access-k6grg\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454841 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454850 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9cg6\" (UniqueName: \"kubernetes.io/projected/c94b8e19-9edb-4a82-95a9-5717afa5ae6d-kube-api-access-b9cg6\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454859 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qwp2\" (UniqueName: \"kubernetes.io/projected/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489-kube-api-access-6qwp2\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454867 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll7l8\" (UniqueName: \"kubernetes.io/projected/d77cc176-a7dd-4510-8f10-a5cc227ca06f-kube-api-access-ll7l8\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.454875 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d77cc176-a7dd-4510-8f10-a5cc227ca06f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.715352 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6189-account-create-5fm2f" event={"ID":"c94b8e19-9edb-4a82-95a9-5717afa5ae6d","Type":"ContainerDied","Data":"831d65e5cc0dae3cdedb599d67ee83ce3a03c00b2e6a1ee5a8408fb97706a679"} Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.715396 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="831d65e5cc0dae3cdedb599d67ee83ce3a03c00b2e6a1ee5a8408fb97706a679" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.715398 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6189-account-create-5fm2f" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.716808 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-wkwfz" event={"ID":"7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489","Type":"ContainerDied","Data":"70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9"} Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.716837 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70b99ee280bf30d2879c711fc140ee01b4a20d209b13e1e8db45d8b073ab38d9" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.716910 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-wkwfz" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.718559 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5205-account-create-2qjmw" event={"ID":"8a055662-deb7-4f6a-bef1-a69e71f155e0","Type":"ContainerDied","Data":"622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1"} Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.718625 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5205-account-create-2qjmw" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.718644 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622d2e726a4b835f37133bb406e4272edbcc4c8d6da0702c88ed956d429ecfe1" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.722283 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-z5stn" event={"ID":"d77cc176-a7dd-4510-8f10-a5cc227ca06f","Type":"ContainerDied","Data":"c3251b191bec65f3d53d8ab08196b981049c2d47881cc49f02915f25d7408bc1"} Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.722321 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3251b191bec65f3d53d8ab08196b981049c2d47881cc49f02915f25d7408bc1" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.722355 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-z5stn" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.723818 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8446-account-create-kqdzx" event={"ID":"dd54ba95-cb56-4068-90c9-06cfd72111c4","Type":"ContainerDied","Data":"ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2"} Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.723853 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab5d52a2162bfeababf81c438dfd804c21c06eeb09aea42032a47fc4388305d2" Nov 24 14:01:56 crc kubenswrapper[4970]: I1124 14:01:56.723893 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8446-account-create-kqdzx" Nov 24 14:01:58 crc kubenswrapper[4970]: I1124 14:01:58.298795 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:01:58 crc kubenswrapper[4970]: I1124 14:01:58.349936 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:01:58 crc kubenswrapper[4970]: I1124 14:01:58.350425 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="dnsmasq-dns" containerID="cri-o://4d3f5f6844380e3314c4ad131865a2361f732989825564ea8b077095cfd2be4f" gracePeriod=10 Nov 24 14:01:58 crc kubenswrapper[4970]: I1124 14:01:58.743688 4970 generic.go:334] "Generic (PLEG): container finished" podID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerID="4d3f5f6844380e3314c4ad131865a2361f732989825564ea8b077095cfd2be4f" exitCode=0 Nov 24 14:01:58 crc kubenswrapper[4970]: I1124 14:01:58.743738 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerDied","Data":"4d3f5f6844380e3314c4ad131865a2361f732989825564ea8b077095cfd2be4f"} Nov 24 14:02:02 crc kubenswrapper[4970]: I1124 14:02:02.229908 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.129336 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.189729 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config\") pod \"95ec9410-e3c8-4d9f-be06-6011987e3c07\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.189876 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb\") pod \"95ec9410-e3c8-4d9f-be06-6011987e3c07\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.190003 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc\") pod \"95ec9410-e3c8-4d9f-be06-6011987e3c07\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.190076 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb\") pod \"95ec9410-e3c8-4d9f-be06-6011987e3c07\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.190221 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkhpf\" (UniqueName: \"kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf\") pod \"95ec9410-e3c8-4d9f-be06-6011987e3c07\" (UID: \"95ec9410-e3c8-4d9f-be06-6011987e3c07\") " Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.195911 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf" (OuterVolumeSpecName: "kube-api-access-mkhpf") pod "95ec9410-e3c8-4d9f-be06-6011987e3c07" (UID: "95ec9410-e3c8-4d9f-be06-6011987e3c07"). InnerVolumeSpecName "kube-api-access-mkhpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.239539 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config" (OuterVolumeSpecName: "config") pod "95ec9410-e3c8-4d9f-be06-6011987e3c07" (UID: "95ec9410-e3c8-4d9f-be06-6011987e3c07"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.242247 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95ec9410-e3c8-4d9f-be06-6011987e3c07" (UID: "95ec9410-e3c8-4d9f-be06-6011987e3c07"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.243537 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95ec9410-e3c8-4d9f-be06-6011987e3c07" (UID: "95ec9410-e3c8-4d9f-be06-6011987e3c07"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.245958 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95ec9410-e3c8-4d9f-be06-6011987e3c07" (UID: "95ec9410-e3c8-4d9f-be06-6011987e3c07"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.291821 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkhpf\" (UniqueName: \"kubernetes.io/projected/95ec9410-e3c8-4d9f-be06-6011987e3c07-kube-api-access-mkhpf\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.291846 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.291856 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.291866 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.291877 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95ec9410-e3c8-4d9f-be06-6011987e3c07-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.803052 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.803480 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf8bcbfcf-bwhb7" event={"ID":"95ec9410-e3c8-4d9f-be06-6011987e3c07","Type":"ContainerDied","Data":"d3a5a81ff75e28e9da579f81d5ba9b77a9b1c6c7053a0075f37f59e3050f1211"} Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.803746 4970 scope.go:117] "RemoveContainer" containerID="4d3f5f6844380e3314c4ad131865a2361f732989825564ea8b077095cfd2be4f" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.818472 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6tvk" event={"ID":"145d1a69-5b39-4ec0-8544-3034bded1955","Type":"ContainerStarted","Data":"e95f90e4f466ba059b65cb676c994ce48521fabee21932369b74146fb40c82c1"} Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.857025 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c6tvk" podStartSLOduration=2.5769235139999997 podStartE2EDuration="12.856993645s" podCreationTimestamp="2025-11-24 14:01:52 +0000 UTC" firstStartedPulling="2025-11-24 14:01:53.604535338 +0000 UTC m=+928.892292631" lastFinishedPulling="2025-11-24 14:02:03.884605469 +0000 UTC m=+939.172362762" observedRunningTime="2025-11-24 14:02:04.840278849 +0000 UTC m=+940.128036182" watchObservedRunningTime="2025-11-24 14:02:04.856993645 +0000 UTC m=+940.144750978" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.922888 4970 scope.go:117] "RemoveContainer" containerID="d14f13fd9417d4d135eb70de7827fda658c8ddb1e927290180c252bf25cf6a91" Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.924961 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:02:04 crc kubenswrapper[4970]: I1124 14:02:04.934899 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf8bcbfcf-bwhb7"] Nov 24 14:02:05 crc kubenswrapper[4970]: I1124 14:02:05.488827 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" path="/var/lib/kubelet/pods/95ec9410-e3c8-4d9f-be06-6011987e3c07/volumes" Nov 24 14:02:09 crc kubenswrapper[4970]: I1124 14:02:09.888702 4970 generic.go:334] "Generic (PLEG): container finished" podID="145d1a69-5b39-4ec0-8544-3034bded1955" containerID="e95f90e4f466ba059b65cb676c994ce48521fabee21932369b74146fb40c82c1" exitCode=0 Nov 24 14:02:09 crc kubenswrapper[4970]: I1124 14:02:09.888839 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6tvk" event={"ID":"145d1a69-5b39-4ec0-8544-3034bded1955","Type":"ContainerDied","Data":"e95f90e4f466ba059b65cb676c994ce48521fabee21932369b74146fb40c82c1"} Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.191554 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.204859 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.204964 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.205043 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.206218 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.206350 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c" gracePeriod=600 Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.310750 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle\") pod \"145d1a69-5b39-4ec0-8544-3034bded1955\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.310841 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r685f\" (UniqueName: \"kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f\") pod \"145d1a69-5b39-4ec0-8544-3034bded1955\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.310888 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data\") pod \"145d1a69-5b39-4ec0-8544-3034bded1955\" (UID: \"145d1a69-5b39-4ec0-8544-3034bded1955\") " Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.325979 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f" (OuterVolumeSpecName: "kube-api-access-r685f") pod "145d1a69-5b39-4ec0-8544-3034bded1955" (UID: "145d1a69-5b39-4ec0-8544-3034bded1955"). InnerVolumeSpecName "kube-api-access-r685f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.342027 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "145d1a69-5b39-4ec0-8544-3034bded1955" (UID: "145d1a69-5b39-4ec0-8544-3034bded1955"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.384689 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data" (OuterVolumeSpecName: "config-data") pod "145d1a69-5b39-4ec0-8544-3034bded1955" (UID: "145d1a69-5b39-4ec0-8544-3034bded1955"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.413246 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.413291 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/145d1a69-5b39-4ec0-8544-3034bded1955-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.413305 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r685f\" (UniqueName: \"kubernetes.io/projected/145d1a69-5b39-4ec0-8544-3034bded1955-kube-api-access-r685f\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.905965 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c" exitCode=0 Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.906180 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c"} Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.906683 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e"} Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.906769 4970 scope.go:117] "RemoveContainer" containerID="4c03893179ad6d3f79ad48c2d2f782400686aee8f1162838694406d3477c4099" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.909293 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c6tvk" event={"ID":"145d1a69-5b39-4ec0-8544-3034bded1955","Type":"ContainerDied","Data":"b392c5b213095248b217435c5067684d1a3919b44ca5beba0e6349b924f6eded"} Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.909324 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b392c5b213095248b217435c5067684d1a3919b44ca5beba0e6349b924f6eded" Nov 24 14:02:11 crc kubenswrapper[4970]: I1124 14:02:11.909373 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c6tvk" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206057 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206699 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206712 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206725 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a055662-deb7-4f6a-bef1-a69e71f155e0" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206732 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a055662-deb7-4f6a-bef1-a69e71f155e0" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206744 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd54ba95-cb56-4068-90c9-06cfd72111c4" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206750 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd54ba95-cb56-4068-90c9-06cfd72111c4" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206761 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="dnsmasq-dns" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206767 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="dnsmasq-dns" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206776 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94b8e19-9edb-4a82-95a9-5717afa5ae6d" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206781 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94b8e19-9edb-4a82-95a9-5717afa5ae6d" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206792 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206797 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206811 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145d1a69-5b39-4ec0-8544-3034bded1955" containerName="keystone-db-sync" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206816 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="145d1a69-5b39-4ec0-8544-3034bded1955" containerName="keystone-db-sync" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206827 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77cc176-a7dd-4510-8f10-a5cc227ca06f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206833 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77cc176-a7dd-4510-8f10-a5cc227ca06f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: E1124 14:02:12.206844 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="init" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206850 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="init" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.206987 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94b8e19-9edb-4a82-95a9-5717afa5ae6d" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207008 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a055662-deb7-4f6a-bef1-a69e71f155e0" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207022 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207034 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77cc176-a7dd-4510-8f10-a5cc227ca06f" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207051 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="95ec9410-e3c8-4d9f-be06-6011987e3c07" containerName="dnsmasq-dns" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207063 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" containerName="mariadb-database-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207074 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="145d1a69-5b39-4ec0-8544-3034bded1955" containerName="keystone-db-sync" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207084 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd54ba95-cb56-4068-90c9-06cfd72111c4" containerName="mariadb-account-create" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.207873 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.226396 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.278635 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-lhz5z"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.279654 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.294039 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.294239 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.294315 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.294348 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.294796 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ptrd5" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.332970 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333040 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333068 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333089 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333109 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7szdn\" (UniqueName: \"kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333132 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333153 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333177 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b69mw\" (UniqueName: \"kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333213 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333237 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333277 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.333313 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.349397 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lhz5z"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433783 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433834 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433865 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433882 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433896 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433912 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7szdn\" (UniqueName: \"kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433932 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433949 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433968 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b69mw\" (UniqueName: \"kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.433998 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.434020 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.434053 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.437686 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.439622 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.440783 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.441692 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.451123 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.451283 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.451777 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.453389 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.469244 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.474088 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.474944 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b69mw\" (UniqueName: \"kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw\") pod \"dnsmasq-dns-5cd88b5f5c-wvsvd\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.502266 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7szdn\" (UniqueName: \"kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn\") pod \"keystone-bootstrap-lhz5z\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.524615 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.586224 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jnshr"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.587567 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.603804 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.603993 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8drnv" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.607368 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.619880 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.620456 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jnshr"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.653852 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.655477 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.664660 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.664830 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.664928 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.672114 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-tft92" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.687034 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738540 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738616 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4t5b\" (UniqueName: \"kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738648 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738666 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738709 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7js\" (UniqueName: \"kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738748 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738824 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.738862 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.766693 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-nmb4b"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.768118 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.774281 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.790351 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sxnp5" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.809424 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nmb4b"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.820548 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.825174 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.830530 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.830934 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.844658 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-dt7wf"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.845767 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.846998 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847036 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4t5b\" (UniqueName: \"kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847064 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847087 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847113 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847141 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7js\" (UniqueName: \"kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847165 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847217 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847251 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847281 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.847303 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jrtn\" (UniqueName: \"kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.848224 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.849528 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.849668 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.851792 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.858925 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.860226 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rf4sz" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.860413 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.860563 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.860771 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.862218 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.879649 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dt7wf"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.886440 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4t5b\" (UniqueName: \"kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b\") pod \"neutron-db-sync-jnshr\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.891277 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7js\" (UniqueName: \"kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js\") pod \"horizon-d6c797b4f-hv87q\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.935125 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.939210 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954783 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954817 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954843 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954869 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jrtn\" (UniqueName: \"kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954891 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954911 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954925 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvn6f\" (UniqueName: \"kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954943 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954958 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg5g6\" (UniqueName: \"kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.954982 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955002 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955027 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955062 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955096 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955115 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.955139 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.958171 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.962265 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.962502 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.962768 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjmj8" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.962891 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.962977 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.963604 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.970479 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.987124 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jrtn\" (UniqueName: \"kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn\") pod \"barbican-db-sync-nmb4b\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:12 crc kubenswrapper[4970]: I1124 14:02:12.987187 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.014561 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.038648 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-6j5rx"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.039754 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.042817 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.043241 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dpq9j" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.043447 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.054913 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.063625 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6j5rx"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066248 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066285 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjqq7\" (UniqueName: \"kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066310 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066338 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066355 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066390 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066418 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066434 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066450 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066466 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066480 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvn6f\" (UniqueName: \"kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066501 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066518 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg5g6\" (UniqueName: \"kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066540 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066556 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066575 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066605 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066625 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066644 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066659 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066695 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.066930 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.067061 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.068193 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.068353 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.074002 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.077919 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.078808 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.083416 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.083812 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.085369 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.087070 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg5g6\" (UniqueName: \"kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.090567 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvn6f\" (UniqueName: \"kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f\") pod \"cinder-db-sync-dt7wf\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.092844 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.092940 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.093593 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.094541 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.102859 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.106325 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.140342 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.147296 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.149379 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.151761 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.151962 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.153962 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196716 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196799 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196832 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196866 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196899 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196927 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr44l\" (UniqueName: \"kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196959 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.196989 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.197038 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.197070 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201353 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkqw\" (UniqueName: \"kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201558 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201623 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201678 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201696 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201734 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjqq7\" (UniqueName: \"kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201782 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201834 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201873 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201901 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201948 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wssvd\" (UniqueName: \"kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.201972 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.202005 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.202036 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.205708 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.206171 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.206453 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.206979 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.207694 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.215268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.215327 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.237943 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.245846 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.258946 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjqq7\" (UniqueName: \"kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.319937 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.319985 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.320022 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkqw\" (UniqueName: \"kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.320043 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.320082 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.320158 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323402 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323475 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323500 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323551 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323593 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323628 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323648 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323680 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323705 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323724 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9xp\" (UniqueName: \"kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323774 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wssvd\" (UniqueName: \"kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323795 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323810 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323828 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323888 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323949 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323978 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr44l\" (UniqueName: \"kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.323993 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.325352 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.326013 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.327651 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.327896 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.330568 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.331532 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.331543 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.342354 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.343308 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wssvd\" (UniqueName: \"kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.344087 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data\") pod \"placement-db-sync-6j5rx\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.349345 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr44l\" (UniqueName: \"kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.362496 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.363844 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.365237 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkqw\" (UniqueName: \"kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw\") pod \"horizon-768b98ddb9-ngqb5\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.365974 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.364796 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc\") pod \"dnsmasq-dns-798745f775-pg5gx\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.372720 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6j5rx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.392013 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.396357 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.421686 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.422877 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426418 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426476 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426515 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9xp\" (UniqueName: \"kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426590 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426615 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426679 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426713 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.426741 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.427470 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.427719 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.432435 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.433830 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.434137 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.441909 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.442143 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.447539 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9xp\" (UniqueName: \"kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.493002 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.560309 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-lhz5z"] Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.590396 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: W1124 14:02:13.666263 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9791d1a9_1c8b_4a2f_ac38_1f169e13dd4a.slice/crio-b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681 WatchSource:0}: Error finding container b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681: Status 404 returned error can't find the container with id b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681 Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.707493 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.778508 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:02:13 crc kubenswrapper[4970]: W1124 14:02:13.789820 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c6bc70_16db_41fd_a10a_192f6a40d306.slice/crio-9beaf39425d07fae76a5208913e684f62462f48f4c6c828964e2aa4304252fae WatchSource:0}: Error finding container 9beaf39425d07fae76a5208913e684f62462f48f4c6c828964e2aa4304252fae: Status 404 returned error can't find the container with id 9beaf39425d07fae76a5208913e684f62462f48f4c6c828964e2aa4304252fae Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.984070 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerStarted","Data":"9beaf39425d07fae76a5208913e684f62462f48f4c6c828964e2aa4304252fae"} Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.994724 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-nmb4b"] Nov 24 14:02:13 crc kubenswrapper[4970]: W1124 14:02:13.996122 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48c12423_0ed8_4c46_bfa2_249b9c67eb14.slice/crio-cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189 WatchSource:0}: Error finding container cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189: Status 404 returned error can't find the container with id cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189 Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.996984 4970 generic.go:334] "Generic (PLEG): container finished" podID="9f935427-b40f-4659-8b12-53ee756aafc0" containerID="79f707deeecfdd6067eaf009e9e5de2caa92628d8883e45630c475b3315f654e" exitCode=0 Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.997054 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" event={"ID":"9f935427-b40f-4659-8b12-53ee756aafc0","Type":"ContainerDied","Data":"79f707deeecfdd6067eaf009e9e5de2caa92628d8883e45630c475b3315f654e"} Nov 24 14:02:13 crc kubenswrapper[4970]: I1124 14:02:13.997082 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" event={"ID":"9f935427-b40f-4659-8b12-53ee756aafc0","Type":"ContainerStarted","Data":"12b47047a8b93c326cdba78193597deb89d0d84f0600017b2f801c742fd22e48"} Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.002938 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.012770 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lhz5z" event={"ID":"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a","Type":"ContainerStarted","Data":"b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681"} Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.049395 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jnshr"] Nov 24 14:02:14 crc kubenswrapper[4970]: W1124 14:02:14.073311 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d9908e3_7a2c_4649_af58_9f4d24c5f06c.slice/crio-a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22 WatchSource:0}: Error finding container a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22: Status 404 returned error can't find the container with id a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22 Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.268169 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-dt7wf"] Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.400736 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:14 crc kubenswrapper[4970]: W1124 14:02:14.438085 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e744296_fb2c_4a6d_9ccd_bdfe324c6f0e.slice/crio-018d99d97f8b34c4cba4603aaeee7ee46f4abd3d394641fb3a7603bc11c13660 WatchSource:0}: Error finding container 018d99d97f8b34c4cba4603aaeee7ee46f4abd3d394641fb3a7603bc11c13660: Status 404 returned error can't find the container with id 018d99d97f8b34c4cba4603aaeee7ee46f4abd3d394641fb3a7603bc11c13660 Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.496373 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:14 crc kubenswrapper[4970]: W1124 14:02:14.504953 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e9cccb9_0d4a_42cc_83d3_10fc7cb20709.slice/crio-36305872e763c2b5e5e8bb58cf88e37c346e4de8e7f82f61edfb018478fed0c8 WatchSource:0}: Error finding container 36305872e763c2b5e5e8bb58cf88e37c346e4de8e7f82f61edfb018478fed0c8: Status 404 returned error can't find the container with id 36305872e763c2b5e5e8bb58cf88e37c346e4de8e7f82f61edfb018478fed0c8 Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.553994 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.568774 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-6j5rx"] Nov 24 14:02:14 crc kubenswrapper[4970]: W1124 14:02:14.593553 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod461c95e6_1aaf_44cc_b090_430a4379e671.slice/crio-d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2 WatchSource:0}: Error finding container d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2: Status 404 returned error can't find the container with id d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2 Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.598178 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685193 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b69mw\" (UniqueName: \"kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685246 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685361 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685409 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685448 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.685472 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc\") pod \"9f935427-b40f-4659-8b12-53ee756aafc0\" (UID: \"9f935427-b40f-4659-8b12-53ee756aafc0\") " Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.693121 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw" (OuterVolumeSpecName: "kube-api-access-b69mw") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "kube-api-access-b69mw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.714505 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.720404 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.727523 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.734104 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.754445 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.760050 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config" (OuterVolumeSpecName: "config") pod "9f935427-b40f-4659-8b12-53ee756aafc0" (UID: "9f935427-b40f-4659-8b12-53ee756aafc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788569 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b69mw\" (UniqueName: \"kubernetes.io/projected/9f935427-b40f-4659-8b12-53ee756aafc0-kube-api-access-b69mw\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788830 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788843 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788855 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788865 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:14 crc kubenswrapper[4970]: I1124 14:02:14.788873 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f935427-b40f-4659-8b12-53ee756aafc0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.057742 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jnshr" event={"ID":"6d9908e3-7a2c-4649-af58-9f4d24c5f06c","Type":"ContainerStarted","Data":"a274e436d6160821373d877d910076c422db0eda9a63de18d0b30e4acecef05d"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.057798 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jnshr" event={"ID":"6d9908e3-7a2c-4649-af58-9f4d24c5f06c","Type":"ContainerStarted","Data":"a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.071498 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" event={"ID":"9f935427-b40f-4659-8b12-53ee756aafc0","Type":"ContainerDied","Data":"12b47047a8b93c326cdba78193597deb89d0d84f0600017b2f801c742fd22e48"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.071557 4970 scope.go:117] "RemoveContainer" containerID="79f707deeecfdd6067eaf009e9e5de2caa92628d8883e45630c475b3315f654e" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.071796 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd88b5f5c-wvsvd" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.077630 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jnshr" podStartSLOduration=3.077612657 podStartE2EDuration="3.077612657s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:15.076883178 +0000 UTC m=+950.364640471" watchObservedRunningTime="2025-11-24 14:02:15.077612657 +0000 UTC m=+950.365369950" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.083209 4970 generic.go:334] "Generic (PLEG): container finished" podID="d585ac61-e10b-428a-93d5-bb49e436b423" containerID="05cd5f3f6f279bd6b69b1fa81dadbf57a7a3814bd9663e9ad94e87210ec762fb" exitCode=0 Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.083265 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798745f775-pg5gx" event={"ID":"d585ac61-e10b-428a-93d5-bb49e436b423","Type":"ContainerDied","Data":"05cd5f3f6f279bd6b69b1fa81dadbf57a7a3814bd9663e9ad94e87210ec762fb"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.083289 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798745f775-pg5gx" event={"ID":"d585ac61-e10b-428a-93d5-bb49e436b423","Type":"ContainerStarted","Data":"65456d31ac74b655b9449166c6f43ce28d302ef2823b3bcbc03843273fb58c9d"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.091953 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nmb4b" event={"ID":"48c12423-0ed8-4c46-bfa2-249b9c67eb14","Type":"ContainerStarted","Data":"cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.099533 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dt7wf" event={"ID":"f85ae910-cbec-47ff-9f7e-bba2545cbcae","Type":"ContainerStarted","Data":"aa0cb720d0634e4b6cdb9d5ebfe809f4dc4dbb1829476b59aa9a51ebb4203a04"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.112119 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-768b98ddb9-ngqb5" event={"ID":"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e","Type":"ContainerStarted","Data":"018d99d97f8b34c4cba4603aaeee7ee46f4abd3d394641fb3a7603bc11c13660"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.116924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerStarted","Data":"36305872e763c2b5e5e8bb58cf88e37c346e4de8e7f82f61edfb018478fed0c8"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.168411 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerStarted","Data":"c833beefa71401e5b951e400afd753dae82ac8bf8329b185f3f444c76dc9ea5f"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.178853 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6j5rx" event={"ID":"461c95e6-1aaf-44cc-b090-430a4379e671","Type":"ContainerStarted","Data":"d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.187926 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lhz5z" event={"ID":"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a","Type":"ContainerStarted","Data":"8ffdb7fd15ddb66c3c218b48044bfd5556055b39b03a73c6674eb36af583b0a2"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.259463 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerStarted","Data":"d0b2aac2c31bef2795d236ba11c76b1db6dd58d27b652d0c63d923bbf798e9ec"} Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.331291 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.365449 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd88b5f5c-wvsvd"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.395125 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-lhz5z" podStartSLOduration=3.395097843 podStartE2EDuration="3.395097843s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:15.215818015 +0000 UTC m=+950.503575308" watchObservedRunningTime="2025-11-24 14:02:15.395097843 +0000 UTC m=+950.682855136" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.443910 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.452643 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.456996 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.467644 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.527097 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f935427-b40f-4659-8b12-53ee756aafc0" path="/var/lib/kubelet/pods/9f935427-b40f-4659-8b12-53ee756aafc0/volumes" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.527980 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:15 crc kubenswrapper[4970]: E1124 14:02:15.528426 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f935427-b40f-4659-8b12-53ee756aafc0" containerName="init" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.528508 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f935427-b40f-4659-8b12-53ee756aafc0" containerName="init" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.528765 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f935427-b40f-4659-8b12-53ee756aafc0" containerName="init" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.529799 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.529967 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.642088 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.642159 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.642197 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.642242 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfqbb\" (UniqueName: \"kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.642267 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.743419 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfqbb\" (UniqueName: \"kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.743468 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.743570 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.743615 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.743640 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.744017 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.744492 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.747439 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.751773 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.760721 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfqbb\" (UniqueName: \"kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb\") pod \"horizon-6b65d8c45c-62sxx\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:15 crc kubenswrapper[4970]: I1124 14:02:15.900856 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.338099 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798745f775-pg5gx" event={"ID":"d585ac61-e10b-428a-93d5-bb49e436b423","Type":"ContainerStarted","Data":"01e4067254d35945e8ce39dae1528124b870157c014f906231235a1586bc650b"} Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.338450 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.343526 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerStarted","Data":"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a"} Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.403648 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerStarted","Data":"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718"} Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.405408 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798745f775-pg5gx" podStartSLOduration=4.405385655 podStartE2EDuration="4.405385655s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:16.382273834 +0000 UTC m=+951.670031127" watchObservedRunningTime="2025-11-24 14:02:16.405385655 +0000 UTC m=+951.693142948" Nov 24 14:02:16 crc kubenswrapper[4970]: I1124 14:02:16.577769 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.464188 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerStarted","Data":"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992"} Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.464715 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-log" containerID="cri-o://bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" gracePeriod=30 Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.464771 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-httpd" containerID="cri-o://d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" gracePeriod=30 Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.472247 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-log" containerID="cri-o://6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" gracePeriod=30 Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.472744 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-httpd" containerID="cri-o://9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" gracePeriod=30 Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.530987 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.53095519 podStartE2EDuration="5.53095519s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:17.503108625 +0000 UTC m=+952.790865918" watchObservedRunningTime="2025-11-24 14:02:17.53095519 +0000 UTC m=+952.818712483" Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.531218 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerStarted","Data":"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f"} Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.531278 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b65d8c45c-62sxx" event={"ID":"45352dac-3fb0-42b1-ab54-48e454a65972","Type":"ContainerStarted","Data":"c945c25a29d9c4f6ea37bd9f0687fc3ae71d7214a0498e11333ecd316834ba6f"} Nov 24 14:02:17 crc kubenswrapper[4970]: I1124 14:02:17.538069 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.538055184 podStartE2EDuration="4.538055184s" podCreationTimestamp="2025-11-24 14:02:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:17.525124128 +0000 UTC m=+952.812881421" watchObservedRunningTime="2025-11-24 14:02:17.538055184 +0000 UTC m=+952.825812477" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.224242 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.232473 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.304755 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.304824 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.304856 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9xp\" (UniqueName: \"kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.304931 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.304961 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305001 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305035 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305064 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjqq7\" (UniqueName: \"kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305092 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305176 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305211 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305243 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305295 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305323 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305383 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs\") pod \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\" (UID: \"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305427 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs\") pod \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\" (UID: \"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709\") " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.305875 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.310862 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.311727 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs" (OuterVolumeSpecName: "logs") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.311775 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs" (OuterVolumeSpecName: "logs") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.314006 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.314271 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts" (OuterVolumeSpecName: "scripts") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.316788 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.317395 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts" (OuterVolumeSpecName: "scripts") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.319724 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7" (OuterVolumeSpecName: "kube-api-access-kjqq7") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "kube-api-access-kjqq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.329027 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp" (OuterVolumeSpecName: "kube-api-access-7q9xp") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "kube-api-access-7q9xp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.363640 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.378070 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.385887 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.389388 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.397858 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data" (OuterVolumeSpecName: "config-data") pod "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" (UID: "6e9cccb9-0d4a-42cc-83d3-10fc7cb20709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.401924 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data" (OuterVolumeSpecName: "config-data") pod "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" (UID: "533365f6-3e2b-4fb1-a591-4edb9fd9d8a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410787 4970 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410822 4970 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410832 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410845 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410853 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9xp\" (UniqueName: \"kubernetes.io/projected/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-kube-api-access-7q9xp\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410890 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410907 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410919 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410928 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410937 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjqq7\" (UniqueName: \"kubernetes.io/projected/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-kube-api-access-kjqq7\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410947 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410958 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410967 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410977 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410987 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.410998 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.430027 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.432020 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.486132 4970 generic.go:334] "Generic (PLEG): container finished" podID="9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" containerID="8ffdb7fd15ddb66c3c218b48044bfd5556055b39b03a73c6674eb36af583b0a2" exitCode=0 Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.486194 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lhz5z" event={"ID":"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a","Type":"ContainerDied","Data":"8ffdb7fd15ddb66c3c218b48044bfd5556055b39b03a73c6674eb36af583b0a2"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490666 4970 generic.go:334] "Generic (PLEG): container finished" podID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerID="d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" exitCode=0 Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490688 4970 generic.go:334] "Generic (PLEG): container finished" podID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerID="bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" exitCode=143 Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490730 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerDied","Data":"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490748 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerDied","Data":"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490758 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6e9cccb9-0d4a-42cc-83d3-10fc7cb20709","Type":"ContainerDied","Data":"36305872e763c2b5e5e8bb58cf88e37c346e4de8e7f82f61edfb018478fed0c8"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490771 4970 scope.go:117] "RemoveContainer" containerID="d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.490851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505005 4970 generic.go:334] "Generic (PLEG): container finished" podID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerID="9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" exitCode=0 Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505033 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerDied","Data":"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505065 4970 generic.go:334] "Generic (PLEG): container finished" podID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerID="6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" exitCode=143 Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505078 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerDied","Data":"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505088 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"533365f6-3e2b-4fb1-a591-4edb9fd9d8a2","Type":"ContainerDied","Data":"c833beefa71401e5b951e400afd753dae82ac8bf8329b185f3f444c76dc9ea5f"} Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.505019 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.511885 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.511905 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.543609 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.573331 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.602827 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: E1124 14:02:18.603329 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603348 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: E1124 14:02:18.603382 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603391 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: E1124 14:02:18.603415 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603423 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: E1124 14:02:18.603438 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603446 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603676 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603694 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603705 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-log" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.603730 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" containerName="glance-httpd" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.605067 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.613617 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.621180 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-kjmj8" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.621230 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.621478 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.622291 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.625920 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.640337 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.654732 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.656399 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.658634 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.660099 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.680466 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719433 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719511 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719540 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719556 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz9dr\" (UniqueName: \"kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719589 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719631 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719664 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719678 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719695 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhr8\" (UniqueName: \"kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719719 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719738 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719763 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719785 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719815 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719836 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.719864 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822513 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822557 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz9dr\" (UniqueName: \"kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822593 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822611 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822636 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822650 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822671 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhr8\" (UniqueName: \"kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822697 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822716 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822745 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822769 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822802 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822824 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822853 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822874 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.822912 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.823348 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.823590 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.823752 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.824046 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.824060 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.824304 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.830858 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.831731 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.844240 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.846852 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.847689 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.856088 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.858353 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.858972 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhr8\" (UniqueName: \"kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.863111 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.865279 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz9dr\" (UniqueName: \"kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.913764 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.933770 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.974236 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:18 crc kubenswrapper[4970]: I1124 14:02:18.981901 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:19 crc kubenswrapper[4970]: I1124 14:02:19.492833 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="533365f6-3e2b-4fb1-a591-4edb9fd9d8a2" path="/var/lib/kubelet/pods/533365f6-3e2b-4fb1-a591-4edb9fd9d8a2/volumes" Nov 24 14:02:19 crc kubenswrapper[4970]: I1124 14:02:19.493751 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e9cccb9-0d4a-42cc-83d3-10fc7cb20709" path="/var/lib/kubelet/pods/6e9cccb9-0d4a-42cc-83d3-10fc7cb20709/volumes" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.319792 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.350088 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.352409 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.357908 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364721 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364780 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364805 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364828 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364876 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364904 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx7xb\" (UniqueName: \"kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.364950 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.396645 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.433613 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466127 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466190 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466215 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466238 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466283 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466309 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx7xb\" (UniqueName: \"kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466353 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.466436 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.467343 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.468922 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.469334 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.477505 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.485843 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.493323 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.493363 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bb8f9fccd-2m5xk"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.495589 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.495851 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.515124 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx7xb\" (UniqueName: \"kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb\") pod \"horizon-c5f4c5bbd-xvgxp\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.516408 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bb8f9fccd-2m5xk"] Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.670952 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-tls-certs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671048 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-scripts\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671131 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp7lp\" (UniqueName: \"kubernetes.io/projected/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-kube-api-access-fp7lp\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671161 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-combined-ca-bundle\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671195 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-secret-key\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671214 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-logs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.671233 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-config-data\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.694872 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.772935 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp7lp\" (UniqueName: \"kubernetes.io/projected/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-kube-api-access-fp7lp\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.772994 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-combined-ca-bundle\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.773039 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-secret-key\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.773068 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-logs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.773099 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-config-data\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.773159 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-tls-certs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.773205 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-scripts\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.774079 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-scripts\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.775309 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-logs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.778165 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-config-data\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.778185 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-combined-ca-bundle\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.781449 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-tls-certs\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.782910 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-horizon-secret-key\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.791098 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp7lp\" (UniqueName: \"kubernetes.io/projected/070dfeee-2ee7-4b9e-bdbf-7320a48a0659-kube-api-access-fp7lp\") pod \"horizon-5bb8f9fccd-2m5xk\" (UID: \"070dfeee-2ee7-4b9e-bdbf-7320a48a0659\") " pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:21 crc kubenswrapper[4970]: I1124 14:02:21.876559 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:23 crc kubenswrapper[4970]: I1124 14:02:23.397756 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:23 crc kubenswrapper[4970]: I1124 14:02:23.448132 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:02:23 crc kubenswrapper[4970]: I1124 14:02:23.448440 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" containerID="cri-o://1308215d54a7e659b575a86127109b865bd51732edbe2ae8c3e9a5534ad45258" gracePeriod=10 Nov 24 14:02:24 crc kubenswrapper[4970]: I1124 14:02:24.571657 4970 generic.go:334] "Generic (PLEG): container finished" podID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerID="1308215d54a7e659b575a86127109b865bd51732edbe2ae8c3e9a5534ad45258" exitCode=0 Nov 24 14:02:24 crc kubenswrapper[4970]: I1124 14:02:24.571739 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" event={"ID":"e2bba80e-8ca2-4b12-b404-cac67c1a69e6","Type":"ContainerDied","Data":"1308215d54a7e659b575a86127109b865bd51732edbe2ae8c3e9a5534ad45258"} Nov 24 14:02:28 crc kubenswrapper[4970]: I1124 14:02:28.297998 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Nov 24 14:02:31 crc kubenswrapper[4970]: E1124 14:02:31.932866 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645" Nov 24 14:02:31 crc kubenswrapper[4970]: E1124 14:02:31.933433 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8jrtn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-nmb4b_openstack(48c12423-0ed8-4c46-bfa2-249b9c67eb14): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:02:31 crc kubenswrapper[4970]: E1124 14:02:31.934638 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-nmb4b" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" Nov 24 14:02:31 crc kubenswrapper[4970]: I1124 14:02:31.984717 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085287 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085383 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085475 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085516 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085553 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.085572 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7szdn\" (UniqueName: \"kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn\") pod \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\" (UID: \"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a\") " Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.090715 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.091464 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.102613 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn" (OuterVolumeSpecName: "kube-api-access-7szdn") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "kube-api-access-7szdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.102845 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts" (OuterVolumeSpecName: "scripts") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.111810 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.117250 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data" (OuterVolumeSpecName: "config-data") pod "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" (UID: "9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187890 4970 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187922 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187931 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187941 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187952 4970 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.187960 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7szdn\" (UniqueName: \"kubernetes.io/projected/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a-kube-api-access-7szdn\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.639001 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-lhz5z" Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.639021 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-lhz5z" event={"ID":"9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a","Type":"ContainerDied","Data":"b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681"} Nov 24 14:02:32 crc kubenswrapper[4970]: I1124 14:02:32.639054 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1103ebc2621e826c4e72d92ae94765cdf767ec43c9870bc5f727e944db38681" Nov 24 14:02:32 crc kubenswrapper[4970]: E1124 14:02:32.641601 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645\\\"\"" pod="openstack/barbican-db-sync-nmb4b" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.101604 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-lhz5z"] Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.109246 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-lhz5z"] Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.144032 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-7x8kz"] Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.144558 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" containerName="keystone-bootstrap" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.144594 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" containerName="keystone-bootstrap" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.144801 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" containerName="keystone-bootstrap" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.145564 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.148861 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ptrd5" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.149224 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.149528 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.149662 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.149866 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.154808 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7x8kz"] Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.297940 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.306879 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw2hk\" (UniqueName: \"kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.306977 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.307025 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.307045 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.307075 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.307121 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.408901 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw2hk\" (UniqueName: \"kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.408980 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.409017 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.409034 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.409057 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.409094 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.412797 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.413721 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.413827 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.414178 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.416510 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.425284 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw2hk\" (UniqueName: \"kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk\") pod \"keystone-bootstrap-7x8kz\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.476980 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:33 crc kubenswrapper[4970]: I1124 14:02:33.489073 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a" path="/var/lib/kubelet/pods/9791d1a9-1c8b-4a2f-ac38-1f169e13dd4a/volumes" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.743730 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.743922 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wssvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-6j5rx_openstack(461c95e6-1aaf-44cc-b090-430a4379e671): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.745117 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-6j5rx" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.770597 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.773419 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n576h5cfh5f9h554h648hc4h648h54dhch5bch57h649h587h546hd6h9ch8ch67h65h7ch75h595h5bh6fh7dh57bhfdh565h586h9h5f8h5c5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jfqbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6b65d8c45c-62sxx_openstack(45352dac-3fb0-42b1-ab54-48e454a65972): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.777908 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-6b65d8c45c-62sxx" podUID="45352dac-3fb0-42b1-ab54-48e454a65972" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.786116 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.786295 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf9h66bh94h55h78hb6h674h5cchc8h7fh579h5c7hc4hffh68fh5f5h6bhf8h66fh685h64bh78h678h5fh679h59fh5d9h5dh59ch667h58ch9dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffkqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-768b98ddb9-ngqb5_openstack(6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:02:33 crc kubenswrapper[4970]: E1124 14:02:33.789148 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:65b94ff9fcd486845fb0544583bf2a973246a61a0ad32340fb92d632285f1057\\\"\"]" pod="openstack/horizon-768b98ddb9-ngqb5" podUID="6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" Nov 24 14:02:34 crc kubenswrapper[4970]: E1124 14:02:34.657241 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:7dd2e0dbb6bb5a6cecd1763e43479ca8cb6a0c502534e83c8795c0da2b50e099\\\"\"" pod="openstack/placement-db-sync-6j5rx" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.103225 4970 scope.go:117] "RemoveContainer" containerID="bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.228053 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.234132 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.243459 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.265800 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts\") pod \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266268 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266307 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266348 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffkqw\" (UniqueName: \"kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw\") pod \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266393 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data\") pod \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266439 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv826\" (UniqueName: \"kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.266482 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.267843 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts" (OuterVolumeSpecName: "scripts") pod "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" (UID: "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.268450 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data" (OuterVolumeSpecName: "config-data") pod "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" (UID: "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.271094 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw" (OuterVolumeSpecName: "kube-api-access-ffkqw") pod "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" (UID: "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e"). InnerVolumeSpecName "kube-api-access-ffkqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.277857 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826" (OuterVolumeSpecName: "kube-api-access-rv826") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "kube-api-access-rv826". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.345539 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config" (OuterVolumeSpecName: "config") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.347454 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.348718 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368790 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs\") pod \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368836 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfqbb\" (UniqueName: \"kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb\") pod \"45352dac-3fb0-42b1-ab54-48e454a65972\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368856 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368877 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key\") pod \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\" (UID: \"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368916 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb\") pod \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\" (UID: \"e2bba80e-8ca2-4b12-b404-cac67c1a69e6\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368940 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data\") pod \"45352dac-3fb0-42b1-ab54-48e454a65972\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.368976 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts\") pod \"45352dac-3fb0-42b1-ab54-48e454a65972\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369011 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs\") pod \"45352dac-3fb0-42b1-ab54-48e454a65972\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369041 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key\") pod \"45352dac-3fb0-42b1-ab54-48e454a65972\" (UID: \"45352dac-3fb0-42b1-ab54-48e454a65972\") " Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369101 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs" (OuterVolumeSpecName: "logs") pod "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" (UID: "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369371 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv826\" (UniqueName: \"kubernetes.io/projected/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-kube-api-access-rv826\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369394 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369404 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369412 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369420 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369430 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369439 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffkqw\" (UniqueName: \"kubernetes.io/projected/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-kube-api-access-ffkqw\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369448 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.369867 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts" (OuterVolumeSpecName: "scripts") pod "45352dac-3fb0-42b1-ab54-48e454a65972" (UID: "45352dac-3fb0-42b1-ab54-48e454a65972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.370319 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs" (OuterVolumeSpecName: "logs") pod "45352dac-3fb0-42b1-ab54-48e454a65972" (UID: "45352dac-3fb0-42b1-ab54-48e454a65972"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.370453 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data" (OuterVolumeSpecName: "config-data") pod "45352dac-3fb0-42b1-ab54-48e454a65972" (UID: "45352dac-3fb0-42b1-ab54-48e454a65972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.372273 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "45352dac-3fb0-42b1-ab54-48e454a65972" (UID: "45352dac-3fb0-42b1-ab54-48e454a65972"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.372875 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" (UID: "6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.374061 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb" (OuterVolumeSpecName: "kube-api-access-jfqbb") pod "45352dac-3fb0-42b1-ab54-48e454a65972" (UID: "45352dac-3fb0-42b1-ab54-48e454a65972"). InnerVolumeSpecName "kube-api-access-jfqbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.414432 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.418596 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e2bba80e-8ca2-4b12-b404-cac67c1a69e6" (UID: "e2bba80e-8ca2-4b12-b404-cac67c1a69e6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471658 4970 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/45352dac-3fb0-42b1-ab54-48e454a65972-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471688 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfqbb\" (UniqueName: \"kubernetes.io/projected/45352dac-3fb0-42b1-ab54-48e454a65972-kube-api-access-jfqbb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471699 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471708 4970 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471717 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e2bba80e-8ca2-4b12-b404-cac67c1a69e6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471724 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471733 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/45352dac-3fb0-42b1-ab54-48e454a65972-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.471742 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45352dac-3fb0-42b1-ab54-48e454a65972-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.729650 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b65d8c45c-62sxx" event={"ID":"45352dac-3fb0-42b1-ab54-48e454a65972","Type":"ContainerDied","Data":"c945c25a29d9c4f6ea37bd9f0687fc3ae71d7214a0498e11333ecd316834ba6f"} Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.729696 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b65d8c45c-62sxx" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.736381 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" event={"ID":"e2bba80e-8ca2-4b12-b404-cac67c1a69e6","Type":"ContainerDied","Data":"2ee88eef8fcfc1a9d7d4c73e31ecdaa0c4fd7f9558729cdc5bc5caccf8697e0c"} Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.736401 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.737834 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-768b98ddb9-ngqb5" Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.737834 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-768b98ddb9-ngqb5" event={"ID":"6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e","Type":"ContainerDied","Data":"018d99d97f8b34c4cba4603aaeee7ee46f4abd3d394641fb3a7603bc11c13660"} Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.808984 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.816649 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b65d8c45c-62sxx"] Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.826771 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.838176 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb6b7c77c-5jz68"] Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.853380 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:42 crc kubenswrapper[4970]: I1124 14:02:42.861080 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-768b98ddb9-ngqb5"] Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.297313 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cb6b7c77c-5jz68" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.316136 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.316375 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lvn6f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-dt7wf_openstack(f85ae910-cbec-47ff-9f7e-bba2545cbcae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.318074 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-dt7wf" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.336974 4970 scope.go:117] "RemoveContainer" containerID="d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.337272 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992\": container with ID starting with d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992 not found: ID does not exist" containerID="d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337300 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992"} err="failed to get container status \"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992\": rpc error: code = NotFound desc = could not find container \"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992\": container with ID starting with d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992 not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337321 4970 scope.go:117] "RemoveContainer" containerID="bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.337496 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718\": container with ID starting with bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718 not found: ID does not exist" containerID="bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337524 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718"} err="failed to get container status \"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718\": rpc error: code = NotFound desc = could not find container \"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718\": container with ID starting with bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718 not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337538 4970 scope.go:117] "RemoveContainer" containerID="d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337718 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992"} err="failed to get container status \"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992\": rpc error: code = NotFound desc = could not find container \"d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992\": container with ID starting with d560c02c3de8e2c6e5baab14abe6512d0f85563f2a694d4530338713a69db992 not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337732 4970 scope.go:117] "RemoveContainer" containerID="bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337959 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718"} err="failed to get container status \"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718\": rpc error: code = NotFound desc = could not find container \"bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718\": container with ID starting with bf85c1703d2b9b9e07c2370cb780af38eb3b75f727b0e41c6d28d909471f7718 not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.337990 4970 scope.go:117] "RemoveContainer" containerID="9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.400684 4970 scope.go:117] "RemoveContainer" containerID="6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.456291 4970 scope.go:117] "RemoveContainer" containerID="9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.457529 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f\": container with ID starting with 9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f not found: ID does not exist" containerID="9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.457694 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f"} err="failed to get container status \"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f\": rpc error: code = NotFound desc = could not find container \"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f\": container with ID starting with 9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.457729 4970 scope.go:117] "RemoveContainer" containerID="6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.458354 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a\": container with ID starting with 6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a not found: ID does not exist" containerID="6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.458398 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a"} err="failed to get container status \"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a\": rpc error: code = NotFound desc = could not find container \"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a\": container with ID starting with 6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.458418 4970 scope.go:117] "RemoveContainer" containerID="9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.458769 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f"} err="failed to get container status \"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f\": rpc error: code = NotFound desc = could not find container \"9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f\": container with ID starting with 9b8c4f0c6245a9d98c802ccacbb5af25c3e6287484d7c6afb7541f2e3ecaa23f not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.458836 4970 scope.go:117] "RemoveContainer" containerID="6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.462204 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a"} err="failed to get container status \"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a\": rpc error: code = NotFound desc = could not find container \"6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a\": container with ID starting with 6f1b08aa28d9db7ea55bb626464ba51206ea0786ae565bfb40a442051fd0c96a not found: ID does not exist" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.462252 4970 scope.go:117] "RemoveContainer" containerID="1308215d54a7e659b575a86127109b865bd51732edbe2ae8c3e9a5534ad45258" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.500892 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45352dac-3fb0-42b1-ab54-48e454a65972" path="/var/lib/kubelet/pods/45352dac-3fb0-42b1-ab54-48e454a65972/volumes" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.501748 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e" path="/var/lib/kubelet/pods/6e744296-fb2c-4a6d-9ccd-bdfe324c6f0e/volumes" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.502262 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" path="/var/lib/kubelet/pods/e2bba80e-8ca2-4b12-b404-cac67c1a69e6/volumes" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.527922 4970 scope.go:117] "RemoveContainer" containerID="8ec867edc0e42e24e0847720796430851695c4457627c6cbb2755a34da91f59a" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.759304 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerStarted","Data":"4c25755885db79faff32cf45456d3cf5b3037c98ea5b4eeb387b3fdcb411cd6d"} Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.761265 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerStarted","Data":"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8"} Nov 24 14:02:43 crc kubenswrapper[4970]: E1124 14:02:43.766267 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-dt7wf" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.935741 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bb8f9fccd-2m5xk"] Nov 24 14:02:43 crc kubenswrapper[4970]: I1124 14:02:43.942700 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-7x8kz"] Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.023950 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.193290 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.774289 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerStarted","Data":"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.774534 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerStarted","Data":"e9739c067d4d25dae1d3fd0136afe4671383c3e2d64f9c651aa1fadaea0e8ff6"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.777270 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7x8kz" event={"ID":"75d29428-7adf-4a22-b656-36c9b931c8d4","Type":"ContainerStarted","Data":"7086366a4a9317eb1c139ef54322b201ed90d9007bfa270c3b24f2292892c37f"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.777300 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7x8kz" event={"ID":"75d29428-7adf-4a22-b656-36c9b931c8d4","Type":"ContainerStarted","Data":"81cb21b120af8945ef6d42ca793a08b078066b396e70668dea4e945eb5c4955a"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.794401 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-7x8kz" podStartSLOduration=11.794381726 podStartE2EDuration="11.794381726s" podCreationTimestamp="2025-11-24 14:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:44.79413252 +0000 UTC m=+980.081889833" watchObservedRunningTime="2025-11-24 14:02:44.794381726 +0000 UTC m=+980.082139019" Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.794729 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerStarted","Data":"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.794859 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d6c797b4f-hv87q" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon-log" containerID="cri-o://5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" gracePeriod=30 Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.794988 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d6c797b4f-hv87q" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon" containerID="cri-o://8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" gracePeriod=30 Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.796703 4970 generic.go:334] "Generic (PLEG): container finished" podID="6d9908e3-7a2c-4649-af58-9f4d24c5f06c" containerID="a274e436d6160821373d877d910076c422db0eda9a63de18d0b30e4acecef05d" exitCode=0 Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.796769 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jnshr" event={"ID":"6d9908e3-7a2c-4649-af58-9f4d24c5f06c","Type":"ContainerDied","Data":"a274e436d6160821373d877d910076c422db0eda9a63de18d0b30e4acecef05d"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.800626 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bb8f9fccd-2m5xk" event={"ID":"070dfeee-2ee7-4b9e-bdbf-7320a48a0659","Type":"ContainerStarted","Data":"0f537a1ad32d6d6f47d40040c635acf42ca401b555eaf05102e5d631b0668742"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.800666 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bb8f9fccd-2m5xk" event={"ID":"070dfeee-2ee7-4b9e-bdbf-7320a48a0659","Type":"ContainerStarted","Data":"001eb4efe83d08b6b0c265136cc4f40ed83c397dd9f3fc6a2c794138620f9c80"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.800675 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bb8f9fccd-2m5xk" event={"ID":"070dfeee-2ee7-4b9e-bdbf-7320a48a0659","Type":"ContainerStarted","Data":"61b7c0d5ab023ac59b5499f5f5254f8999e7589d09cdebfcac6882c200cf8038"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.802622 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerStarted","Data":"ca522efedbf9731e67c7b3e65f8e485ba5ff868b422ba102b15ddfee9e01717d"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.802852 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerStarted","Data":"056363c1e0d4e437d7a1045c2498d225b2ec2f0473410e76f98de4d4e9493418"} Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.816053 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d6c797b4f-hv87q" podStartSLOduration=4.468543332 podStartE2EDuration="32.816033769s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="2025-11-24 14:02:13.795685972 +0000 UTC m=+949.083443265" lastFinishedPulling="2025-11-24 14:02:42.143176399 +0000 UTC m=+977.430933702" observedRunningTime="2025-11-24 14:02:44.812534227 +0000 UTC m=+980.100291540" watchObservedRunningTime="2025-11-24 14:02:44.816033769 +0000 UTC m=+980.103791052" Nov 24 14:02:44 crc kubenswrapper[4970]: I1124 14:02:44.866289 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bb8f9fccd-2m5xk" podStartSLOduration=23.866267431 podStartE2EDuration="23.866267431s" podCreationTimestamp="2025-11-24 14:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:44.860279153 +0000 UTC m=+980.148036456" watchObservedRunningTime="2025-11-24 14:02:44.866267431 +0000 UTC m=+980.154024724" Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.159663 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.812480 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerStarted","Data":"dfc762669e70a4b4cb8fd4eeb37addccd129f4996b2fb624ce881f0dda8fa73e"} Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.813112 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerStarted","Data":"925d3907d7fea59114ea1af77789b632b7a81ba9d31370bd348568492e37dd2f"} Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.815701 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerStarted","Data":"6f5a1170f608d84dfefcb3970f5b308afb8b70329afa1e99dcc81638e534855d"} Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.819878 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerStarted","Data":"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e"} Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.820040 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-log" containerID="cri-o://450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" gracePeriod=30 Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.820215 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-httpd" containerID="cri-o://111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" gracePeriod=30 Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.823249 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerStarted","Data":"626642014a83a07d9d403f33cf0bb77a9f601135a762f85ddf2d43b2cae5b431"} Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.879501 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c5f4c5bbd-xvgxp" podStartSLOduration=24.879479389 podStartE2EDuration="24.879479389s" podCreationTimestamp="2025-11-24 14:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:45.841876172 +0000 UTC m=+981.129633465" watchObservedRunningTime="2025-11-24 14:02:45.879479389 +0000 UTC m=+981.167236682" Nov 24 14:02:45 crc kubenswrapper[4970]: I1124 14:02:45.890728 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.890708627 podStartE2EDuration="27.890708627s" podCreationTimestamp="2025-11-24 14:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:45.87651336 +0000 UTC m=+981.164270663" watchObservedRunningTime="2025-11-24 14:02:45.890708627 +0000 UTC m=+981.178465910" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.240434 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.365829 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config\") pod \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.366380 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle\") pod \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.366518 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4t5b\" (UniqueName: \"kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b\") pod \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\" (UID: \"6d9908e3-7a2c-4649-af58-9f4d24c5f06c\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.385732 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b" (OuterVolumeSpecName: "kube-api-access-q4t5b") pod "6d9908e3-7a2c-4649-af58-9f4d24c5f06c" (UID: "6d9908e3-7a2c-4649-af58-9f4d24c5f06c"). InnerVolumeSpecName "kube-api-access-q4t5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.402658 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config" (OuterVolumeSpecName: "config") pod "6d9908e3-7a2c-4649-af58-9f4d24c5f06c" (UID: "6d9908e3-7a2c-4649-af58-9f4d24c5f06c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.468535 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4t5b\" (UniqueName: \"kubernetes.io/projected/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-kube-api-access-q4t5b\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.468556 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.515711 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d9908e3-7a2c-4649-af58-9f4d24c5f06c" (UID: "6d9908e3-7a2c-4649-af58-9f4d24c5f06c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.570042 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9908e3-7a2c-4649-af58-9f4d24c5f06c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.583934 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773396 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773444 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773516 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773534 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773566 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fhr8\" (UniqueName: \"kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773670 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773722 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.773775 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle\") pod \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\" (UID: \"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716\") " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.774008 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs" (OuterVolumeSpecName: "logs") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.774060 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.774271 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.778307 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts" (OuterVolumeSpecName: "scripts") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.778312 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8" (OuterVolumeSpecName: "kube-api-access-5fhr8") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "kube-api-access-5fhr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.778308 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.802655 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.825964 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data" (OuterVolumeSpecName: "config-data") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.831440 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" (UID: "ec05e3ef-28f1-4b25-a8c1-53a5c73f8716"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.833972 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jnshr" event={"ID":"6d9908e3-7a2c-4649-af58-9f4d24c5f06c","Type":"ContainerDied","Data":"a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22"} Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.834006 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a737061f30879946508001cf3c6ab1d5c97c366c703bb5697845ba296a446c22" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.834053 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jnshr" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.851475 4970 generic.go:334] "Generic (PLEG): container finished" podID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerID="111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" exitCode=0 Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.851521 4970 generic.go:334] "Generic (PLEG): container finished" podID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerID="450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" exitCode=143 Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.851652 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.853084 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerDied","Data":"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e"} Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.853141 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerDied","Data":"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c"} Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.853158 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ec05e3ef-28f1-4b25-a8c1-53a5c73f8716","Type":"ContainerDied","Data":"e9739c067d4d25dae1d3fd0136afe4671383c3e2d64f9c651aa1fadaea0e8ff6"} Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.854512 4970 scope.go:117] "RemoveContainer" containerID="111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.877168 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fhr8\" (UniqueName: \"kubernetes.io/projected/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-kube-api-access-5fhr8\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.877424 4970 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.877485 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.877562 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.877942 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.878019 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.878118 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.896124 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.911757 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.916016 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.926728 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:46 crc kubenswrapper[4970]: E1124 14:02:46.927197 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="init" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927221 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="init" Nov 24 14:02:46 crc kubenswrapper[4970]: E1124 14:02:46.927254 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-log" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927263 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-log" Nov 24 14:02:46 crc kubenswrapper[4970]: E1124 14:02:46.927280 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-httpd" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927289 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-httpd" Nov 24 14:02:46 crc kubenswrapper[4970]: E1124 14:02:46.927312 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9908e3-7a2c-4649-af58-9f4d24c5f06c" containerName="neutron-db-sync" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927320 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9908e3-7a2c-4649-af58-9f4d24c5f06c" containerName="neutron-db-sync" Nov 24 14:02:46 crc kubenswrapper[4970]: E1124 14:02:46.927334 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927342 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927537 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-httpd" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927568 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9908e3-7a2c-4649-af58-9f4d24c5f06c" containerName="neutron-db-sync" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927608 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2bba80e-8ca2-4b12-b404-cac67c1a69e6" containerName="dnsmasq-dns" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.927618 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" containerName="glance-log" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.928853 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.932166 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.932368 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.934713 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:46 crc kubenswrapper[4970]: I1124 14:02:46.981838 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.077867 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.080172 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083409 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083461 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083493 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083512 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083726 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083884 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.083948 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.084039 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zgq5\" (UniqueName: \"kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.085319 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.165390 4970 scope.go:117] "RemoveContainer" containerID="450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186443 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7wj\" (UniqueName: \"kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186498 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186543 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186600 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186653 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zgq5\" (UniqueName: \"kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186703 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186729 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186751 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186790 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186823 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186854 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186885 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186919 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.186951 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.187657 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.191863 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.193683 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.195923 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.195988 4970 scope.go:117] "RemoveContainer" containerID="111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" Nov 24 14:02:47 crc kubenswrapper[4970]: E1124 14:02:47.196442 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e\": container with ID starting with 111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e not found: ID does not exist" containerID="111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196472 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e"} err="failed to get container status \"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e\": rpc error: code = NotFound desc = could not find container \"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e\": container with ID starting with 111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e not found: ID does not exist" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196493 4970 scope.go:117] "RemoveContainer" containerID="450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196785 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196931 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: E1124 14:02:47.196940 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c\": container with ID starting with 450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c not found: ID does not exist" containerID="450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196971 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c"} err="failed to get container status \"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c\": rpc error: code = NotFound desc = could not find container \"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c\": container with ID starting with 450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c not found: ID does not exist" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.196990 4970 scope.go:117] "RemoveContainer" containerID="111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.197993 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.198109 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e"} err="failed to get container status \"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e\": rpc error: code = NotFound desc = could not find container \"111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e\": container with ID starting with 111a99a6cb2e757fefd132a100c645d237317ed8b82558425f92029b7636d38e not found: ID does not exist" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.198140 4970 scope.go:117] "RemoveContainer" containerID="450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.199845 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c"} err="failed to get container status \"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c\": rpc error: code = NotFound desc = could not find container \"450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c\": container with ID starting with 450c0c447e7c059bd5accb047339a9a05627fc7773d07c23fef6a7ecd5ffb26c not found: ID does not exist" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.225091 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zgq5\" (UniqueName: \"kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.256400 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.257343 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.257865 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.261359 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.261795 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8drnv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.261804 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.261937 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.276705 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288293 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288333 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288396 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288424 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288457 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7wj\" (UniqueName: \"kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.288478 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.289395 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.291544 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.292309 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.292399 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.292458 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.319590 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7wj\" (UniqueName: \"kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj\") pod \"dnsmasq-dns-5b95cfcf9c-7t8hv\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.391390 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bjhd\" (UniqueName: \"kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.391443 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.391474 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.391524 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.391597 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.434048 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.493186 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.493312 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bjhd\" (UniqueName: \"kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.493333 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.493363 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.493416 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.494441 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec05e3ef-28f1-4b25-a8c1-53a5c73f8716" path="/var/lib/kubelet/pods/ec05e3ef-28f1-4b25-a8c1-53a5c73f8716/volumes" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.501750 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.501862 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.502507 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.502793 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.517076 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bjhd\" (UniqueName: \"kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd\") pod \"neutron-66b7fdd9cd-76pz9\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.556962 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.617330 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.875836 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerStarted","Data":"fc83c7a39d41bbe2ffec6d69e5f826a7b84219fcf3063cfca8d35a2f5f218a2d"} Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.876188 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-log" containerID="cri-o://dfc762669e70a4b4cb8fd4eeb37addccd129f4996b2fb624ce881f0dda8fa73e" gracePeriod=30 Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.876354 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-httpd" containerID="cri-o://fc83c7a39d41bbe2ffec6d69e5f826a7b84219fcf3063cfca8d35a2f5f218a2d" gracePeriod=30 Nov 24 14:02:47 crc kubenswrapper[4970]: I1124 14:02:47.903041 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=29.903021897 podStartE2EDuration="29.903021897s" podCreationTimestamp="2025-11-24 14:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:47.895915689 +0000 UTC m=+983.183672992" watchObservedRunningTime="2025-11-24 14:02:47.903021897 +0000 UTC m=+983.190779190" Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.007090 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:02:48 crc kubenswrapper[4970]: W1124 14:02:48.013725 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bcbbd0d_82de_4751_b0ca_aae14c1f1337.slice/crio-d75375e48d12c89e63768fbf221c3a61827566b4d1bf092914c37569cb52e735 WatchSource:0}: Error finding container d75375e48d12c89e63768fbf221c3a61827566b4d1bf092914c37569cb52e735: Status 404 returned error can't find the container with id d75375e48d12c89e63768fbf221c3a61827566b4d1bf092914c37569cb52e735 Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.390703 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.472445 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.945077 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerStarted","Data":"e32bf6d9fdbf0d20ef614eae6800597e898710ee4f305b4b9ead0035d48e7912"} Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.954251 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nmb4b" event={"ID":"48c12423-0ed8-4c46-bfa2-249b9c67eb14","Type":"ContainerStarted","Data":"86867005583fbb6f73a0238863e598f48c3d0a36437cb1fe5458a206f6642b5a"} Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.959875 4970 generic.go:334] "Generic (PLEG): container finished" podID="75d29428-7adf-4a22-b656-36c9b931c8d4" containerID="7086366a4a9317eb1c139ef54322b201ed90d9007bfa270c3b24f2292892c37f" exitCode=0 Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.959980 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7x8kz" event={"ID":"75d29428-7adf-4a22-b656-36c9b931c8d4","Type":"ContainerDied","Data":"7086366a4a9317eb1c139ef54322b201ed90d9007bfa270c3b24f2292892c37f"} Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.982970 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.983014 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:48 crc kubenswrapper[4970]: I1124 14:02:48.987961 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-nmb4b" podStartSLOduration=3.442006947 podStartE2EDuration="36.987947507s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="2025-11-24 14:02:13.998560294 +0000 UTC m=+949.286317587" lastFinishedPulling="2025-11-24 14:02:47.544500854 +0000 UTC m=+982.832258147" observedRunningTime="2025-11-24 14:02:48.982698507 +0000 UTC m=+984.270455800" watchObservedRunningTime="2025-11-24 14:02:48.987947507 +0000 UTC m=+984.275704800" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.011326 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6j5rx" event={"ID":"461c95e6-1aaf-44cc-b090-430a4379e671","Type":"ContainerStarted","Data":"91b7c1a2ff2869740ebfe758ca0357f2e6452f87f228a6b36f63b7b4d6893352"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.023147 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerStarted","Data":"2b231424689eda4ae371e87749a89d5bdd1337f8d159e5622d8787552270d28d"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.030244 4970 generic.go:334] "Generic (PLEG): container finished" podID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerID="569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af" exitCode=0 Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.030467 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" event={"ID":"1bcbbd0d-82de-4751-b0ca-aae14c1f1337","Type":"ContainerDied","Data":"569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.030556 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" event={"ID":"1bcbbd0d-82de-4751-b0ca-aae14c1f1337","Type":"ContainerStarted","Data":"d75375e48d12c89e63768fbf221c3a61827566b4d1bf092914c37569cb52e735"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.033400 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-6j5rx" podStartSLOduration=3.319540789 podStartE2EDuration="37.033376751s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="2025-11-24 14:02:14.59567813 +0000 UTC m=+949.883435423" lastFinishedPulling="2025-11-24 14:02:48.309514092 +0000 UTC m=+983.597271385" observedRunningTime="2025-11-24 14:02:49.027062724 +0000 UTC m=+984.314820017" watchObservedRunningTime="2025-11-24 14:02:49.033376751 +0000 UTC m=+984.321134044" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.042853 4970 generic.go:334] "Generic (PLEG): container finished" podID="59d38929-08d9-4058-bd36-a4985dd32797" containerID="fc83c7a39d41bbe2ffec6d69e5f826a7b84219fcf3063cfca8d35a2f5f218a2d" exitCode=0 Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.042894 4970 generic.go:334] "Generic (PLEG): container finished" podID="59d38929-08d9-4058-bd36-a4985dd32797" containerID="dfc762669e70a4b4cb8fd4eeb37addccd129f4996b2fb624ce881f0dda8fa73e" exitCode=143 Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.042924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerDied","Data":"fc83c7a39d41bbe2ffec6d69e5f826a7b84219fcf3063cfca8d35a2f5f218a2d"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.042956 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerDied","Data":"dfc762669e70a4b4cb8fd4eeb37addccd129f4996b2fb624ce881f0dda8fa73e"} Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.422823 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577055 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577113 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz9dr\" (UniqueName: \"kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577143 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577197 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577251 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577341 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577464 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577492 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"59d38929-08d9-4058-bd36-a4985dd32797\" (UID: \"59d38929-08d9-4058-bd36-a4985dd32797\") " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.577959 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.578068 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.579119 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs" (OuterVolumeSpecName: "logs") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.602049 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts" (OuterVolumeSpecName: "scripts") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.605619 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr" (OuterVolumeSpecName: "kube-api-access-sz9dr") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "kube-api-access-sz9dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.615931 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.654678 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6464bbf59c-l44tb"] Nov 24 14:02:49 crc kubenswrapper[4970]: E1124 14:02:49.655074 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-httpd" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.655090 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-httpd" Nov 24 14:02:49 crc kubenswrapper[4970]: E1124 14:02:49.655118 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-log" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.655125 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-log" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.655280 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-httpd" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.655306 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d38929-08d9-4058-bd36-a4985dd32797" containerName="glance-log" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.656174 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.658092 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6464bbf59c-l44tb"] Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.674938 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.675161 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.682069 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.682295 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.682363 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz9dr\" (UniqueName: \"kubernetes.io/projected/59d38929-08d9-4058-bd36-a4985dd32797-kube-api-access-sz9dr\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.682424 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d38929-08d9-4058-bd36-a4985dd32797-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.722068 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.774634 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790533 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-combined-ca-bundle\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790594 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790621 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-ovndb-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790641 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rp8n\" (UniqueName: \"kubernetes.io/projected/ca5d8b57-826f-4811-a8b4-4f48e3082473-kube-api-access-7rp8n\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790704 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-internal-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790740 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-public-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790761 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-httpd-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790803 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.790814 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.801643 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data" (OuterVolumeSpecName: "config-data") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.811500 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "59d38929-08d9-4058-bd36-a4985dd32797" (UID: "59d38929-08d9-4058-bd36-a4985dd32797"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.891678 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-httpd-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892033 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-combined-ca-bundle\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892071 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892108 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-ovndb-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892136 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rp8n\" (UniqueName: \"kubernetes.io/projected/ca5d8b57-826f-4811-a8b4-4f48e3082473-kube-api-access-7rp8n\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892219 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-internal-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892277 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-public-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892325 4970 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.892337 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d38929-08d9-4058-bd36-a4985dd32797-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.918866 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-internal-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.919259 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.921428 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-ovndb-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.921952 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-combined-ca-bundle\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.922740 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rp8n\" (UniqueName: \"kubernetes.io/projected/ca5d8b57-826f-4811-a8b4-4f48e3082473-kube-api-access-7rp8n\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.941140 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-httpd-config\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:49 crc kubenswrapper[4970]: I1124 14:02:49.997359 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca5d8b57-826f-4811-a8b4-4f48e3082473-public-tls-certs\") pod \"neutron-6464bbf59c-l44tb\" (UID: \"ca5d8b57-826f-4811-a8b4-4f48e3082473\") " pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.055090 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.067364 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"59d38929-08d9-4058-bd36-a4985dd32797","Type":"ContainerDied","Data":"925d3907d7fea59114ea1af77789b632b7a81ba9d31370bd348568492e37dd2f"} Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.067433 4970 scope.go:117] "RemoveContainer" containerID="fc83c7a39d41bbe2ffec6d69e5f826a7b84219fcf3063cfca8d35a2f5f218a2d" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.067612 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.100940 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerStarted","Data":"25d030299d57e4ef5ade75a2cba774943a4477dcb0e7d4d461d152708684cf0f"} Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.147484 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.157846 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.201877 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.203391 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.213778 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.214023 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.216275 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.257810 4970 scope.go:117] "RemoveContainer" containerID="dfc762669e70a4b4cb8fd4eeb37addccd129f4996b2fb624ce881f0dda8fa73e" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309463 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h54mr\" (UniqueName: \"kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309734 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309758 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309795 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309819 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309837 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309877 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.309897 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415499 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h54mr\" (UniqueName: \"kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415588 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415614 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415654 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415676 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415694 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415739 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.415760 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.416299 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.417086 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.417479 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.429110 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.429569 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.436790 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.442264 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h54mr\" (UniqueName: \"kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.453230 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.531233 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.573895 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.619351 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.619479 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.620679 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.620760 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw2hk\" (UniqueName: \"kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.620795 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.620853 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data\") pod \"75d29428-7adf-4a22-b656-36c9b931c8d4\" (UID: \"75d29428-7adf-4a22-b656-36c9b931c8d4\") " Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.623410 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.629261 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk" (OuterVolumeSpecName: "kube-api-access-xw2hk") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "kube-api-access-xw2hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.629585 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.645070 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts" (OuterVolumeSpecName: "scripts") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.652815 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.656225 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data" (OuterVolumeSpecName: "config-data") pod "75d29428-7adf-4a22-b656-36c9b931c8d4" (UID: "75d29428-7adf-4a22-b656-36c9b931c8d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.676512 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722800 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw2hk\" (UniqueName: \"kubernetes.io/projected/75d29428-7adf-4a22-b656-36c9b931c8d4-kube-api-access-xw2hk\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722845 4970 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722858 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722869 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722882 4970 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.722894 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75d29428-7adf-4a22-b656-36c9b931c8d4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:50 crc kubenswrapper[4970]: I1124 14:02:50.954844 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6464bbf59c-l44tb"] Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.113469 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerStarted","Data":"9686e90d1b07bb6994e2237f076190b5e98f06e7872b555e81d33f9e8d2cba37"} Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.116347 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" event={"ID":"1bcbbd0d-82de-4751-b0ca-aae14c1f1337","Type":"ContainerStarted","Data":"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e"} Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.117465 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.134222 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6464bbf59c-l44tb" event={"ID":"ca5d8b57-826f-4811-a8b4-4f48e3082473","Type":"ContainerStarted","Data":"c6e2dac192b8916921a6ae3e9a91f64fae7f18da70da4c9cbab08c3e05743ee1"} Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.146846 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerStarted","Data":"c48df9554a6eac5c7de114da9d4e9d3d33153f4f11c4a691480a39cf0cd01b1d"} Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.152984 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" podStartSLOduration=4.152962446 podStartE2EDuration="4.152962446s" podCreationTimestamp="2025-11-24 14:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:51.147835871 +0000 UTC m=+986.435593164" watchObservedRunningTime="2025-11-24 14:02:51.152962446 +0000 UTC m=+986.440719749" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.158246 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-7x8kz" event={"ID":"75d29428-7adf-4a22-b656-36c9b931c8d4","Type":"ContainerDied","Data":"81cb21b120af8945ef6d42ca793a08b078066b396e70668dea4e945eb5c4955a"} Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.158291 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81cb21b120af8945ef6d42ca793a08b078066b396e70668dea4e945eb5c4955a" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.158354 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-7x8kz" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.199269 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8c975fcc4-8cmzb"] Nov 24 14:02:51 crc kubenswrapper[4970]: E1124 14:02:51.208137 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d29428-7adf-4a22-b656-36c9b931c8d4" containerName="keystone-bootstrap" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.208189 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d29428-7adf-4a22-b656-36c9b931c8d4" containerName="keystone-bootstrap" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.208528 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d29428-7adf-4a22-b656-36c9b931c8d4" containerName="keystone-bootstrap" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.209122 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.212733 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.212916 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.213041 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.213067 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ptrd5" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.213173 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.219222 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.252558 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8c975fcc4-8cmzb"] Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.329444 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.353866 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-credential-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.353952 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-combined-ca-bundle\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.353977 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-public-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.354019 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sfsv\" (UniqueName: \"kubernetes.io/projected/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-kube-api-access-9sfsv\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.354050 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-config-data\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.354084 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-internal-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.354129 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-scripts\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.354156 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-fernet-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.455869 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-credential-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456278 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-combined-ca-bundle\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456313 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-public-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456370 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sfsv\" (UniqueName: \"kubernetes.io/projected/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-kube-api-access-9sfsv\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456414 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-config-data\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456456 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-internal-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456524 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-scripts\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.456561 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-fernet-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.464659 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-fernet-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.465998 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-credential-keys\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.466668 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-combined-ca-bundle\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.470471 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-internal-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.472723 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-public-tls-certs\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.475374 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-config-data\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.478858 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-scripts\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.478905 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sfsv\" (UniqueName: \"kubernetes.io/projected/24ee1a5e-62c4-42f5-999b-27ddf44b33ae-kube-api-access-9sfsv\") pod \"keystone-8c975fcc4-8cmzb\" (UID: \"24ee1a5e-62c4-42f5-999b-27ddf44b33ae\") " pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.494405 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d38929-08d9-4058-bd36-a4985dd32797" path="/var/lib/kubelet/pods/59d38929-08d9-4058-bd36-a4985dd32797/volumes" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.696308 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.697102 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.738521 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.877308 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:51 crc kubenswrapper[4970]: I1124 14:02:51.877383 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:02:52 crc kubenswrapper[4970]: I1124 14:02:52.184969 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerStarted","Data":"18743ef26e78757d652cdf4969aa9a5924a012b8f4ce912e98704ff409f00388"} Nov 24 14:02:52 crc kubenswrapper[4970]: I1124 14:02:52.199990 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6464bbf59c-l44tb" event={"ID":"ca5d8b57-826f-4811-a8b4-4f48e3082473","Type":"ContainerStarted","Data":"80f8ac84ea436e05a133e1b56f371de1ba95a17c3eec453ffb19120ba1d259ff"} Nov 24 14:02:52 crc kubenswrapper[4970]: I1124 14:02:52.200673 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:02:52 crc kubenswrapper[4970]: I1124 14:02:52.231134 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66b7fdd9cd-76pz9" podStartSLOduration=5.231111835 podStartE2EDuration="5.231111835s" podCreationTimestamp="2025-11-24 14:02:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:52.224749097 +0000 UTC m=+987.512506380" watchObservedRunningTime="2025-11-24 14:02:52.231111835 +0000 UTC m=+987.518869128" Nov 24 14:02:53 crc kubenswrapper[4970]: I1124 14:02:53.014825 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:02:56 crc kubenswrapper[4970]: I1124 14:02:56.237699 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6464bbf59c-l44tb" event={"ID":"ca5d8b57-826f-4811-a8b4-4f48e3082473","Type":"ContainerStarted","Data":"f0ba6f6b8f1f01d172671b4595d7790ea8a36e308c51a90fb846d7b57e82cd13"} Nov 24 14:02:56 crc kubenswrapper[4970]: I1124 14:02:56.238297 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:02:56 crc kubenswrapper[4970]: I1124 14:02:56.266916 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6464bbf59c-l44tb" podStartSLOduration=7.266895513 podStartE2EDuration="7.266895513s" podCreationTimestamp="2025-11-24 14:02:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:56.262342391 +0000 UTC m=+991.550099684" watchObservedRunningTime="2025-11-24 14:02:56.266895513 +0000 UTC m=+991.554652806" Nov 24 14:02:56 crc kubenswrapper[4970]: I1124 14:02:56.354187 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8c975fcc4-8cmzb"] Nov 24 14:02:56 crc kubenswrapper[4970]: W1124 14:02:56.424216 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24ee1a5e_62c4_42f5_999b_27ddf44b33ae.slice/crio-291913f771ecff52a41a48cdf71857fbdacf4c1d0897fd6f157dae1e63a8bf96 WatchSource:0}: Error finding container 291913f771ecff52a41a48cdf71857fbdacf4c1d0897fd6f157dae1e63a8bf96: Status 404 returned error can't find the container with id 291913f771ecff52a41a48cdf71857fbdacf4c1d0897fd6f157dae1e63a8bf96 Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.255816 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerStarted","Data":"8569b00245973b27500b9015ba1ec080a6f3671a33e1a590c81ad8c215dfdbf1"} Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.260235 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8c975fcc4-8cmzb" event={"ID":"24ee1a5e-62c4-42f5-999b-27ddf44b33ae","Type":"ContainerStarted","Data":"064017f8251767ad219f8cc7e913618f9dd1f1e71caa70fed6fe65e0a0c3fd8c"} Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.260311 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.260329 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8c975fcc4-8cmzb" event={"ID":"24ee1a5e-62c4-42f5-999b-27ddf44b33ae","Type":"ContainerStarted","Data":"291913f771ecff52a41a48cdf71857fbdacf4c1d0897fd6f157dae1e63a8bf96"} Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.265302 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerStarted","Data":"76848b4bd58a805945b82a0892e7891f7816c7970ca978322f2e7d31e9d94a2b"} Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.268128 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerStarted","Data":"979ff5c26104bb1893c8c5d6480b6e690ba7976efbba797fde0171c34e5f3595"} Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.287345 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8c975fcc4-8cmzb" podStartSLOduration=6.28730237 podStartE2EDuration="6.28730237s" podCreationTimestamp="2025-11-24 14:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:57.285710968 +0000 UTC m=+992.573468261" watchObservedRunningTime="2025-11-24 14:02:57.28730237 +0000 UTC m=+992.575059663" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.309718 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.309697954 podStartE2EDuration="11.309697954s" podCreationTimestamp="2025-11-24 14:02:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:57.306260583 +0000 UTC m=+992.594017876" watchObservedRunningTime="2025-11-24 14:02:57.309697954 +0000 UTC m=+992.597455247" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.436779 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.517355 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.517552 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-798745f775-pg5gx" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="dnsmasq-dns" containerID="cri-o://01e4067254d35945e8ce39dae1528124b870157c014f906231235a1586bc650b" gracePeriod=10 Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.558886 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.559715 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.595002 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:02:57 crc kubenswrapper[4970]: I1124 14:02:57.625389 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.286026 4970 generic.go:334] "Generic (PLEG): container finished" podID="d585ac61-e10b-428a-93d5-bb49e436b423" containerID="01e4067254d35945e8ce39dae1528124b870157c014f906231235a1586bc650b" exitCode=0 Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.286092 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798745f775-pg5gx" event={"ID":"d585ac61-e10b-428a-93d5-bb49e436b423","Type":"ContainerDied","Data":"01e4067254d35945e8ce39dae1528124b870157c014f906231235a1586bc650b"} Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.288766 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dt7wf" event={"ID":"f85ae910-cbec-47ff-9f7e-bba2545cbcae","Type":"ContainerStarted","Data":"8c686f47662db039f8847333c14d316dd2516b1f6d03ef268cc37f5cd11be59b"} Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.298870 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerStarted","Data":"162e41175b6f5d6fb2f700370057743c2e9dbb35c5f9e3ab38d4c2b4c1e2e410"} Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.299546 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.299590 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.331027 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-dt7wf" podStartSLOduration=4.201015236 podStartE2EDuration="46.331006036s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="2025-11-24 14:02:14.333685518 +0000 UTC m=+949.621442811" lastFinishedPulling="2025-11-24 14:02:56.463676318 +0000 UTC m=+991.751433611" observedRunningTime="2025-11-24 14:02:58.312098805 +0000 UTC m=+993.599856098" watchObservedRunningTime="2025-11-24 14:02:58.331006036 +0000 UTC m=+993.618763349" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.349991 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.349970058 podStartE2EDuration="8.349970058s" podCreationTimestamp="2025-11-24 14:02:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:02:58.341260688 +0000 UTC m=+993.629017981" watchObservedRunningTime="2025-11-24 14:02:58.349970058 +0000 UTC m=+993.637727351" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.605568 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.688820 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.688887 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.688958 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr44l\" (UniqueName: \"kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.689039 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.689072 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.689102 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0\") pod \"d585ac61-e10b-428a-93d5-bb49e436b423\" (UID: \"d585ac61-e10b-428a-93d5-bb49e436b423\") " Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.708757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l" (OuterVolumeSpecName: "kube-api-access-jr44l") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "kube-api-access-jr44l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.779424 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.780225 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.790661 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr44l\" (UniqueName: \"kubernetes.io/projected/d585ac61-e10b-428a-93d5-bb49e436b423-kube-api-access-jr44l\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.790694 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.790702 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.801007 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.808042 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config" (OuterVolumeSpecName: "config") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.829043 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d585ac61-e10b-428a-93d5-bb49e436b423" (UID: "d585ac61-e10b-428a-93d5-bb49e436b423"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.892547 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.892610 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:58 crc kubenswrapper[4970]: I1124 14:02:58.892629 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d585ac61-e10b-428a-93d5-bb49e436b423-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.314856 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798745f775-pg5gx" event={"ID":"d585ac61-e10b-428a-93d5-bb49e436b423","Type":"ContainerDied","Data":"65456d31ac74b655b9449166c6f43ce28d302ef2823b3bcbc03843273fb58c9d"} Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.316858 4970 scope.go:117] "RemoveContainer" containerID="01e4067254d35945e8ce39dae1528124b870157c014f906231235a1586bc650b" Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.315851 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798745f775-pg5gx" Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.354843 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.363426 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-798745f775-pg5gx"] Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.363771 4970 scope.go:117] "RemoveContainer" containerID="05cd5f3f6f279bd6b69b1fa81dadbf57a7a3814bd9663e9ad94e87210ec762fb" Nov 24 14:02:59 crc kubenswrapper[4970]: I1124 14:02:59.482716 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" path="/var/lib/kubelet/pods/d585ac61-e10b-428a-93d5-bb49e436b423/volumes" Nov 24 14:03:00 crc kubenswrapper[4970]: I1124 14:03:00.677951 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:00 crc kubenswrapper[4970]: I1124 14:03:00.678295 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:00 crc kubenswrapper[4970]: I1124 14:03:00.702395 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:03:00 crc kubenswrapper[4970]: I1124 14:03:00.731160 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:00 crc kubenswrapper[4970]: I1124 14:03:00.733358 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:01 crc kubenswrapper[4970]: I1124 14:03:01.338302 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:01 crc kubenswrapper[4970]: I1124 14:03:01.338668 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:01 crc kubenswrapper[4970]: I1124 14:03:01.697258 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 14:03:01 crc kubenswrapper[4970]: I1124 14:03:01.878919 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5bb8f9fccd-2m5xk" podUID="070dfeee-2ee7-4b9e-bdbf-7320a48a0659" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Nov 24 14:03:02 crc kubenswrapper[4970]: I1124 14:03:02.360057 4970 generic.go:334] "Generic (PLEG): container finished" podID="461c95e6-1aaf-44cc-b090-430a4379e671" containerID="91b7c1a2ff2869740ebfe758ca0357f2e6452f87f228a6b36f63b7b4d6893352" exitCode=0 Nov 24 14:03:02 crc kubenswrapper[4970]: I1124 14:03:02.361030 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6j5rx" event={"ID":"461c95e6-1aaf-44cc-b090-430a4379e671","Type":"ContainerDied","Data":"91b7c1a2ff2869740ebfe758ca0357f2e6452f87f228a6b36f63b7b4d6893352"} Nov 24 14:03:03 crc kubenswrapper[4970]: I1124 14:03:03.322014 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:03 crc kubenswrapper[4970]: I1124 14:03:03.401651 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-798745f775-pg5gx" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.145:5353: i/o timeout" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.447837 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.672370 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6j5rx" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822089 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle\") pod \"461c95e6-1aaf-44cc-b090-430a4379e671\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822407 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs\") pod \"461c95e6-1aaf-44cc-b090-430a4379e671\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822435 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts\") pod \"461c95e6-1aaf-44cc-b090-430a4379e671\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822487 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wssvd\" (UniqueName: \"kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd\") pod \"461c95e6-1aaf-44cc-b090-430a4379e671\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822673 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data\") pod \"461c95e6-1aaf-44cc-b090-430a4379e671\" (UID: \"461c95e6-1aaf-44cc-b090-430a4379e671\") " Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.822775 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs" (OuterVolumeSpecName: "logs") pod "461c95e6-1aaf-44cc-b090-430a4379e671" (UID: "461c95e6-1aaf-44cc-b090-430a4379e671"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.823112 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/461c95e6-1aaf-44cc-b090-430a4379e671-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.829231 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd" (OuterVolumeSpecName: "kube-api-access-wssvd") pod "461c95e6-1aaf-44cc-b090-430a4379e671" (UID: "461c95e6-1aaf-44cc-b090-430a4379e671"). InnerVolumeSpecName "kube-api-access-wssvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.832668 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts" (OuterVolumeSpecName: "scripts") pod "461c95e6-1aaf-44cc-b090-430a4379e671" (UID: "461c95e6-1aaf-44cc-b090-430a4379e671"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.856680 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data" (OuterVolumeSpecName: "config-data") pod "461c95e6-1aaf-44cc-b090-430a4379e671" (UID: "461c95e6-1aaf-44cc-b090-430a4379e671"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.862087 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "461c95e6-1aaf-44cc-b090-430a4379e671" (UID: "461c95e6-1aaf-44cc-b090-430a4379e671"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.924432 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.924471 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.924482 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/461c95e6-1aaf-44cc-b090-430a4379e671-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:04 crc kubenswrapper[4970]: I1124 14:03:04.924492 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wssvd\" (UniqueName: \"kubernetes.io/projected/461c95e6-1aaf-44cc-b090-430a4379e671-kube-api-access-wssvd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.395146 4970 generic.go:334] "Generic (PLEG): container finished" podID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" containerID="86867005583fbb6f73a0238863e598f48c3d0a36437cb1fe5458a206f6642b5a" exitCode=0 Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.395207 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nmb4b" event={"ID":"48c12423-0ed8-4c46-bfa2-249b9c67eb14","Type":"ContainerDied","Data":"86867005583fbb6f73a0238863e598f48c3d0a36437cb1fe5458a206f6642b5a"} Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.398400 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-6j5rx" event={"ID":"461c95e6-1aaf-44cc-b090-430a4379e671","Type":"ContainerDied","Data":"d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2"} Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.398453 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d050bba4d5672ddd7d5bb5c29e28be5038c1f55880988e55dd2da197b1edd8a2" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.398475 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-6j5rx" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813307 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5cd4546cfb-h57bs"] Nov 24 14:03:05 crc kubenswrapper[4970]: E1124 14:03:05.813694 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="dnsmasq-dns" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813708 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="dnsmasq-dns" Nov 24 14:03:05 crc kubenswrapper[4970]: E1124 14:03:05.813727 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" containerName="placement-db-sync" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813733 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" containerName="placement-db-sync" Nov 24 14:03:05 crc kubenswrapper[4970]: E1124 14:03:05.813744 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="init" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813750 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="init" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813921 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d585ac61-e10b-428a-93d5-bb49e436b423" containerName="dnsmasq-dns" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.813931 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" containerName="placement-db-sync" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.814831 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.823403 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.823599 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.823783 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.823935 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.824119 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dpq9j" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.851137 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cd4546cfb-h57bs"] Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943296 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-combined-ca-bundle\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943361 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-internal-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943391 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-scripts\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943427 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-public-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943492 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-config-data\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943528 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj4b2\" (UniqueName: \"kubernetes.io/projected/12d2adeb-5c77-483a-bb89-63b7a50c4935-kube-api-access-cj4b2\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:05 crc kubenswrapper[4970]: I1124 14:03:05.943549 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d2adeb-5c77-483a-bb89-63b7a50c4935-logs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.045852 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-combined-ca-bundle\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.045917 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-internal-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.045949 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-scripts\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.045998 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-public-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.046087 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-config-data\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.046119 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj4b2\" (UniqueName: \"kubernetes.io/projected/12d2adeb-5c77-483a-bb89-63b7a50c4935-kube-api-access-cj4b2\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.046158 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d2adeb-5c77-483a-bb89-63b7a50c4935-logs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.047374 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12d2adeb-5c77-483a-bb89-63b7a50c4935-logs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.051175 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-scripts\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.055331 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-public-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.060865 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-config-data\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.063556 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-internal-tls-certs\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.063738 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj4b2\" (UniqueName: \"kubernetes.io/projected/12d2adeb-5c77-483a-bb89-63b7a50c4935-kube-api-access-cj4b2\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.078519 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12d2adeb-5c77-483a-bb89-63b7a50c4935-combined-ca-bundle\") pod \"placement-5cd4546cfb-h57bs\" (UID: \"12d2adeb-5c77-483a-bb89-63b7a50c4935\") " pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.140124 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.866875 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.961291 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data\") pod \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.961396 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jrtn\" (UniqueName: \"kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn\") pod \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.961452 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle\") pod \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\" (UID: \"48c12423-0ed8-4c46-bfa2-249b9c67eb14\") " Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.965978 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn" (OuterVolumeSpecName: "kube-api-access-8jrtn") pod "48c12423-0ed8-4c46-bfa2-249b9c67eb14" (UID: "48c12423-0ed8-4c46-bfa2-249b9c67eb14"). InnerVolumeSpecName "kube-api-access-8jrtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.966718 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "48c12423-0ed8-4c46-bfa2-249b9c67eb14" (UID: "48c12423-0ed8-4c46-bfa2-249b9c67eb14"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:06 crc kubenswrapper[4970]: I1124 14:03:06.986215 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48c12423-0ed8-4c46-bfa2-249b9c67eb14" (UID: "48c12423-0ed8-4c46-bfa2-249b9c67eb14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.064441 4970 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.064477 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jrtn\" (UniqueName: \"kubernetes.io/projected/48c12423-0ed8-4c46-bfa2-249b9c67eb14-kube-api-access-8jrtn\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.064488 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48c12423-0ed8-4c46-bfa2-249b9c67eb14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.174042 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cd4546cfb-h57bs"] Nov 24 14:03:07 crc kubenswrapper[4970]: W1124 14:03:07.175412 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12d2adeb_5c77_483a_bb89_63b7a50c4935.slice/crio-7c0271f39a61ac02557ae86b1fe477eefb0cf6ff06b1fd861728ba5c261f47fd WatchSource:0}: Error finding container 7c0271f39a61ac02557ae86b1fe477eefb0cf6ff06b1fd861728ba5c261f47fd: Status 404 returned error can't find the container with id 7c0271f39a61ac02557ae86b1fe477eefb0cf6ff06b1fd861728ba5c261f47fd Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.417738 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-nmb4b" event={"ID":"48c12423-0ed8-4c46-bfa2-249b9c67eb14","Type":"ContainerDied","Data":"cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189"} Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.418000 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc387774aee3b86f01b347adeb48628aa03ccbd79b5d682f12a05a123081a189" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.418053 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-nmb4b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.422630 4970 generic.go:334] "Generic (PLEG): container finished" podID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" containerID="8c686f47662db039f8847333c14d316dd2516b1f6d03ef268cc37f5cd11be59b" exitCode=0 Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.422697 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dt7wf" event={"ID":"f85ae910-cbec-47ff-9f7e-bba2545cbcae","Type":"ContainerDied","Data":"8c686f47662db039f8847333c14d316dd2516b1f6d03ef268cc37f5cd11be59b"} Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427051 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerStarted","Data":"964e9e79c84c9240cbac14421e4dab0b7f67045b9f0a53f0fdbae4c9f2102793"} Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427192 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427204 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-central-agent" containerID="cri-o://4c25755885db79faff32cf45456d3cf5b3037c98ea5b4eeb387b3fdcb411cd6d" gracePeriod=30 Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427244 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="proxy-httpd" containerID="cri-o://964e9e79c84c9240cbac14421e4dab0b7f67045b9f0a53f0fdbae4c9f2102793" gracePeriod=30 Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427226 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="sg-core" containerID="cri-o://979ff5c26104bb1893c8c5d6480b6e690ba7976efbba797fde0171c34e5f3595" gracePeriod=30 Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.427302 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-notification-agent" containerID="cri-o://626642014a83a07d9d403f33cf0bb77a9f601135a762f85ddf2d43b2cae5b431" gracePeriod=30 Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.438280 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd4546cfb-h57bs" event={"ID":"12d2adeb-5c77-483a-bb89-63b7a50c4935","Type":"ContainerStarted","Data":"98fcd27ccdfae24df88392811798ce6f477bde4a8013df91d55149ade3345b30"} Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.438376 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd4546cfb-h57bs" event={"ID":"12d2adeb-5c77-483a-bb89-63b7a50c4935","Type":"ContainerStarted","Data":"7c0271f39a61ac02557ae86b1fe477eefb0cf6ff06b1fd861728ba5c261f47fd"} Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.484355 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.76644207 podStartE2EDuration="55.484338445s" podCreationTimestamp="2025-11-24 14:02:12 +0000 UTC" firstStartedPulling="2025-11-24 14:02:14.013822281 +0000 UTC m=+949.301579574" lastFinishedPulling="2025-11-24 14:03:06.731718656 +0000 UTC m=+1002.019475949" observedRunningTime="2025-11-24 14:03:07.47960579 +0000 UTC m=+1002.767363093" watchObservedRunningTime="2025-11-24 14:03:07.484338445 +0000 UTC m=+1002.772095738" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.600300 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-688cdc6cb6-fkrjb"] Nov 24 14:03:07 crc kubenswrapper[4970]: E1124 14:03:07.601027 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" containerName="barbican-db-sync" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.601146 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" containerName="barbican-db-sync" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.601470 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" containerName="barbican-db-sync" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.602870 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.610634 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.610878 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-sxnp5" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.611393 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.623334 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-85596cc87-lvc2b"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.624846 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.627148 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.650345 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-688cdc6cb6-fkrjb"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675378 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675431 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-combined-ca-bundle\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675485 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-logs\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675519 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data-custom\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675536 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptbjf\" (UniqueName: \"kubernetes.io/projected/df1d70bf-d612-4ef1-a113-6be419867e1d-kube-api-access-ptbjf\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675565 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-combined-ca-bundle\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675603 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk58l\" (UniqueName: \"kubernetes.io/projected/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-kube-api-access-hk58l\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675634 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df1d70bf-d612-4ef1-a113-6be419867e1d-logs\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675656 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.675675 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data-custom\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.680330 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-85596cc87-lvc2b"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.727632 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.729170 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.732778 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780116 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-logs\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780227 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data-custom\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780271 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptbjf\" (UniqueName: \"kubernetes.io/projected/df1d70bf-d612-4ef1-a113-6be419867e1d-kube-api-access-ptbjf\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780310 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780355 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-combined-ca-bundle\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780407 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk58l\" (UniqueName: \"kubernetes.io/projected/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-kube-api-access-hk58l\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780437 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780465 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780503 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780561 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df1d70bf-d612-4ef1-a113-6be419867e1d-logs\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780701 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780741 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data-custom\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780799 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780873 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780920 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftvrv\" (UniqueName: \"kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.780964 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-combined-ca-bundle\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.781921 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-logs\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.782235 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/df1d70bf-d612-4ef1-a113-6be419867e1d-logs\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.787132 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.787136 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data-custom\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.798614 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk58l\" (UniqueName: \"kubernetes.io/projected/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-kube-api-access-hk58l\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.805489 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptbjf\" (UniqueName: \"kubernetes.io/projected/df1d70bf-d612-4ef1-a113-6be419867e1d-kube-api-access-ptbjf\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.807461 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-combined-ca-bundle\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.807800 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-combined-ca-bundle\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.810535 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df1d70bf-d612-4ef1-a113-6be419867e1d-config-data-custom\") pod \"barbican-keystone-listener-688cdc6cb6-fkrjb\" (UID: \"df1d70bf-d612-4ef1-a113-6be419867e1d\") " pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.820531 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17b7d151-f71d-4f02-9d52-4c7ba81ab89e-config-data\") pod \"barbican-worker-85596cc87-lvc2b\" (UID: \"17b7d151-f71d-4f02-9d52-4c7ba81ab89e\") " pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.882087 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884450 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884513 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884534 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884554 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884610 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.884651 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftvrv\" (UniqueName: \"kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.885767 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.886328 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.886979 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.887123 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.889604 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.890390 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.890638 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.895156 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.902117 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftvrv\" (UniqueName: \"kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv\") pod \"dnsmasq-dns-66b66f7449-sf8l6\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.938720 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.968484 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85596cc87-lvc2b" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.987887 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.988270 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.988301 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxck\" (UniqueName: \"kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.988330 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:07 crc kubenswrapper[4970]: I1124 14:03:07.988389 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.048970 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.090502 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.090700 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.090763 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.090793 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxck\" (UniqueName: \"kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.090830 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.093130 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.096993 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.099245 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.099277 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.119539 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxck\" (UniqueName: \"kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck\") pod \"barbican-api-64b6b9c8c6-hdl7c\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.236024 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452264 4970 generic.go:334] "Generic (PLEG): container finished" podID="c409523d-f978-43a4-a8a4-1710d56a313b" containerID="964e9e79c84c9240cbac14421e4dab0b7f67045b9f0a53f0fdbae4c9f2102793" exitCode=0 Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452296 4970 generic.go:334] "Generic (PLEG): container finished" podID="c409523d-f978-43a4-a8a4-1710d56a313b" containerID="979ff5c26104bb1893c8c5d6480b6e690ba7976efbba797fde0171c34e5f3595" exitCode=2 Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452304 4970 generic.go:334] "Generic (PLEG): container finished" podID="c409523d-f978-43a4-a8a4-1710d56a313b" containerID="626642014a83a07d9d403f33cf0bb77a9f601135a762f85ddf2d43b2cae5b431" exitCode=0 Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452313 4970 generic.go:334] "Generic (PLEG): container finished" podID="c409523d-f978-43a4-a8a4-1710d56a313b" containerID="4c25755885db79faff32cf45456d3cf5b3037c98ea5b4eeb387b3fdcb411cd6d" exitCode=0 Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452354 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerDied","Data":"964e9e79c84c9240cbac14421e4dab0b7f67045b9f0a53f0fdbae4c9f2102793"} Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452379 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerDied","Data":"979ff5c26104bb1893c8c5d6480b6e690ba7976efbba797fde0171c34e5f3595"} Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452388 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerDied","Data":"626642014a83a07d9d403f33cf0bb77a9f601135a762f85ddf2d43b2cae5b431"} Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.452400 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerDied","Data":"4c25755885db79faff32cf45456d3cf5b3037c98ea5b4eeb387b3fdcb411cd6d"} Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.454684 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd4546cfb-h57bs" event={"ID":"12d2adeb-5c77-483a-bb89-63b7a50c4935","Type":"ContainerStarted","Data":"c3005c29da980c7cf937844f37064b8b6c4a2b6e1b811adc6ab42649ac2ed701"} Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.454709 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.454722 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.489417 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5cd4546cfb-h57bs" podStartSLOduration=3.489394726 podStartE2EDuration="3.489394726s" podCreationTimestamp="2025-11-24 14:03:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:08.471229995 +0000 UTC m=+1003.758987288" watchObservedRunningTime="2025-11-24 14:03:08.489394726 +0000 UTC m=+1003.777152019" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.546530 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-85596cc87-lvc2b"] Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.554334 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-688cdc6cb6-fkrjb"] Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.656694 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.689476 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722206 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722561 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722628 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722768 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg5g6\" (UniqueName: \"kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722810 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722844 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.722888 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts\") pod \"c409523d-f978-43a4-a8a4-1710d56a313b\" (UID: \"c409523d-f978-43a4-a8a4-1710d56a313b\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.726247 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.727362 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts" (OuterVolumeSpecName: "scripts") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.727888 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.731723 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6" (OuterVolumeSpecName: "kube-api-access-sg5g6") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "kube-api-access-sg5g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.751173 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.824993 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.825023 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg5g6\" (UniqueName: \"kubernetes.io/projected/c409523d-f978-43a4-a8a4-1710d56a313b-kube-api-access-sg5g6\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.825038 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c409523d-f978-43a4-a8a4-1710d56a313b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.825051 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.825063 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.833760 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.838434 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data" (OuterVolumeSpecName: "config-data") pod "c409523d-f978-43a4-a8a4-1710d56a313b" (UID: "c409523d-f978-43a4-a8a4-1710d56a313b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.881143 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926236 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926282 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926308 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926418 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926442 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926469 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvn6f\" (UniqueName: \"kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f\") pod \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\" (UID: \"f85ae910-cbec-47ff-9f7e-bba2545cbcae\") " Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926876 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.926891 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c409523d-f978-43a4-a8a4-1710d56a313b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.927026 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.931663 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.932076 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts" (OuterVolumeSpecName: "scripts") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.932094 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f" (OuterVolumeSpecName: "kube-api-access-lvn6f") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "kube-api-access-lvn6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.956895 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:08 crc kubenswrapper[4970]: I1124 14:03:08.962405 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.013743 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data" (OuterVolumeSpecName: "config-data") pod "f85ae910-cbec-47ff-9f7e-bba2545cbcae" (UID: "f85ae910-cbec-47ff-9f7e-bba2545cbcae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028271 4970 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f85ae910-cbec-47ff-9f7e-bba2545cbcae-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028303 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028316 4970 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028325 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028333 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f85ae910-cbec-47ff-9f7e-bba2545cbcae-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.028341 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvn6f\" (UniqueName: \"kubernetes.io/projected/f85ae910-cbec-47ff-9f7e-bba2545cbcae-kube-api-access-lvn6f\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:09 crc kubenswrapper[4970]: W1124 14:03:09.035802 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ab96cbb_2ec7_4e63_8772_6be2ec119b79.slice/crio-52d415d8336b1ffbe87854e337814c68bee3de0d99c00311c047f093e5438067 WatchSource:0}: Error finding container 52d415d8336b1ffbe87854e337814c68bee3de0d99c00311c047f093e5438067: Status 404 returned error can't find the container with id 52d415d8336b1ffbe87854e337814c68bee3de0d99c00311c047f093e5438067 Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.476395 4970 generic.go:334] "Generic (PLEG): container finished" podID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerID="d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18" exitCode=0 Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.509676 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-dt7wf" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516307 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" event={"ID":"3a5b5b3a-6ec0-4454-9a80-0143dd049780","Type":"ContainerDied","Data":"d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516361 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" event={"ID":"3a5b5b3a-6ec0-4454-9a80-0143dd049780","Type":"ContainerStarted","Data":"97990ddfacbfa144f3163ca751a3b371fcc62c7fb7b65ce9508cf57d8852f212"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516477 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" event={"ID":"df1d70bf-d612-4ef1-a113-6be419867e1d","Type":"ContainerStarted","Data":"74e658d31b9d1a96bdb80a52a7df140c020115733e0d8cd6d4fba3fcb5bd7d3a"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516504 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85596cc87-lvc2b" event={"ID":"17b7d151-f71d-4f02-9d52-4c7ba81ab89e","Type":"ContainerStarted","Data":"fe6c313bc661c3bc39dbbdc1197d678d558b247aa1e6730ecd001e2a7de83ebe"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516522 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerStarted","Data":"532b30d02f415d46f0a49cf7392567509233fd9aaec005abeab067ebdf632840"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516537 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerStarted","Data":"52d415d8336b1ffbe87854e337814c68bee3de0d99c00311c047f093e5438067"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516600 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-dt7wf" event={"ID":"f85ae910-cbec-47ff-9f7e-bba2545cbcae","Type":"ContainerDied","Data":"aa0cb720d0634e4b6cdb9d5ebfe809f4dc4dbb1829476b59aa9a51ebb4203a04"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.516621 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa0cb720d0634e4b6cdb9d5ebfe809f4dc4dbb1829476b59aa9a51ebb4203a04" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.521026 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c409523d-f978-43a4-a8a4-1710d56a313b","Type":"ContainerDied","Data":"d0b2aac2c31bef2795d236ba11c76b1db6dd58d27b652d0c63d923bbf798e9ec"} Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.521067 4970 scope.go:117] "RemoveContainer" containerID="964e9e79c84c9240cbac14421e4dab0b7f67045b9f0a53f0fdbae4c9f2102793" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.521099 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.604626 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.619991 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641335 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: E1124 14:03:09.641684 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" containerName="cinder-db-sync" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641696 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" containerName="cinder-db-sync" Nov 24 14:03:09 crc kubenswrapper[4970]: E1124 14:03:09.641723 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="proxy-httpd" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641729 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="proxy-httpd" Nov 24 14:03:09 crc kubenswrapper[4970]: E1124 14:03:09.641740 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="sg-core" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641746 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="sg-core" Nov 24 14:03:09 crc kubenswrapper[4970]: E1124 14:03:09.641763 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-notification-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641768 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-notification-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: E1124 14:03:09.641777 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-central-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641783 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-central-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641950 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="proxy-httpd" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641962 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-notification-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641978 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="ceilometer-central-agent" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.641989 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" containerName="sg-core" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.642004 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" containerName="cinder-db-sync" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.644447 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.650637 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.650828 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.680638 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.684275 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.685989 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.692974 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-rf4sz" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.693272 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.693505 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.693762 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750181 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750284 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s565z\" (UniqueName: \"kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750353 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750441 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750510 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750568 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750598 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750666 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750693 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.750737 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.753469 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.753703 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.754433 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jbc\" (UniqueName: \"kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.827085 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.847000 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.855495 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857779 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857822 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857842 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857861 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857887 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857933 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jbc\" (UniqueName: \"kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857962 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.857986 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s565z\" (UniqueName: \"kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858008 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858038 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858063 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858083 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858096 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.858890 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.859286 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.860284 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.861287 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.861723 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.869896 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.870436 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.871176 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.871194 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.871210 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.872757 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.879255 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.883445 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.889259 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s565z\" (UniqueName: \"kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z\") pod \"cinder-scheduler-0\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.897401 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jbc\" (UniqueName: \"kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc\") pod \"ceilometer-0\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " pod="openstack/ceilometer-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.916566 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.918358 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.925915 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959117 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959407 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959432 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959454 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959477 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959499 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959514 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959539 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5gbd\" (UniqueName: \"kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959559 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959611 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959638 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p2cz\" (UniqueName: \"kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959828 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.959929 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.970623 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:09 crc kubenswrapper[4970]: I1124 14:03:09.998947 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.028950 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.061961 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062047 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062092 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062178 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062200 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062227 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062292 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062487 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062763 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062786 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062819 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5gbd\" (UniqueName: \"kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062839 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062903 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.062946 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p2cz\" (UniqueName: \"kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.063536 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.064241 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.064472 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.065083 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.065335 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.066607 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.076332 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.076398 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.077554 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.079228 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.082516 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p2cz\" (UniqueName: \"kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz\") pod \"cinder-api-0\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.090465 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5gbd\" (UniqueName: \"kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd\") pod \"dnsmasq-dns-7965876c4f-fnzbv\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.216945 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.253541 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.534725 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerStarted","Data":"159b202c0d4d83b56216172cee771df85c777db30392ab8e1cafb700255dab60"} Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.535290 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.561115 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" podStartSLOduration=3.561092762 podStartE2EDuration="3.561092762s" podCreationTimestamp="2025-11-24 14:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:10.549293618 +0000 UTC m=+1005.837050911" watchObservedRunningTime="2025-11-24 14:03:10.561092762 +0000 UTC m=+1005.848850055" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.774335 4970 scope.go:117] "RemoveContainer" containerID="979ff5c26104bb1893c8c5d6480b6e690ba7976efbba797fde0171c34e5f3595" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.878490 4970 scope.go:117] "RemoveContainer" containerID="626642014a83a07d9d403f33cf0bb77a9f601135a762f85ddf2d43b2cae5b431" Nov 24 14:03:10 crc kubenswrapper[4970]: I1124 14:03:10.958673 4970 scope.go:117] "RemoveContainer" containerID="4c25755885db79faff32cf45456d3cf5b3037c98ea5b4eeb387b3fdcb411cd6d" Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.222848 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:11 crc kubenswrapper[4970]: W1124 14:03:11.233613 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cfd8b77_9869_4672_865e_e39e8a752567.slice/crio-52b600c1a4e9932b2e359340548bbf049b4b4cf0c5810af6f752f1f8a1a3b8b2 WatchSource:0}: Error finding container 52b600c1a4e9932b2e359340548bbf049b4b4cf0c5810af6f752f1f8a1a3b8b2: Status 404 returned error can't find the container with id 52b600c1a4e9932b2e359340548bbf049b4b4cf0c5810af6f752f1f8a1a3b8b2 Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.357686 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:11 crc kubenswrapper[4970]: W1124 14:03:11.369282 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70c613a6_e813_420a_9fc3_2b993ac73787.slice/crio-31a83a94e797fe68fee9a338ad091cb4068cd8b49fe6398ddf9986d7160cecce WatchSource:0}: Error finding container 31a83a94e797fe68fee9a338ad091cb4068cd8b49fe6398ddf9986d7160cecce: Status 404 returned error can't find the container with id 31a83a94e797fe68fee9a338ad091cb4068cd8b49fe6398ddf9986d7160cecce Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.497845 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c409523d-f978-43a4-a8a4-1710d56a313b" path="/var/lib/kubelet/pods/c409523d-f978-43a4-a8a4-1710d56a313b/volumes" Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.502190 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.506668 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.547541 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerStarted","Data":"52b600c1a4e9932b2e359340548bbf049b4b4cf0c5810af6f752f1f8a1a3b8b2"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.556105 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" event={"ID":"3a5b5b3a-6ec0-4454-9a80-0143dd049780","Type":"ContainerStarted","Data":"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.556379 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="dnsmasq-dns" containerID="cri-o://377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d" gracePeriod=10 Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.564668 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" event={"ID":"df1d70bf-d612-4ef1-a113-6be419867e1d","Type":"ContainerStarted","Data":"c7d67d50e62d8be62506518331796ec684d9c20f8c5e83d0cd2836c81384e9bc"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.572714 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" event={"ID":"eb87ce82-23c4-4610-a607-c648b7c88b81","Type":"ContainerStarted","Data":"947b7a79252a07fc3896e23d26f0ef4cce595f8dc977c469fe7b37f714bbf543"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.581828 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85596cc87-lvc2b" event={"ID":"17b7d151-f71d-4f02-9d52-4c7ba81ab89e","Type":"ContainerStarted","Data":"21695b08f06f6b4086e1605a5d8ec1db45a3d84455bb5161ab3d40ee46885bad"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.586325 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerStarted","Data":"6b32fcb5e9c8880188200d97b154c76c75b481f3159e447f2a34bc26c224f1b5"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.588822 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerStarted","Data":"31a83a94e797fe68fee9a338ad091cb4068cd8b49fe6398ddf9986d7160cecce"} Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.588902 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:11 crc kubenswrapper[4970]: I1124 14:03:11.591047 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" podStartSLOduration=4.591033702 podStartE2EDuration="4.591033702s" podCreationTimestamp="2025-11-24 14:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:11.584344165 +0000 UTC m=+1006.872101448" watchObservedRunningTime="2025-11-24 14:03:11.591033702 +0000 UTC m=+1006.878790995" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.181625 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215100 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215384 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215486 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215562 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215608 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftvrv\" (UniqueName: \"kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.215640 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc\") pod \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\" (UID: \"3a5b5b3a-6ec0-4454-9a80-0143dd049780\") " Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.263219 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv" (OuterVolumeSpecName: "kube-api-access-ftvrv") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "kube-api-access-ftvrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.317676 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftvrv\" (UniqueName: \"kubernetes.io/projected/3a5b5b3a-6ec0-4454-9a80-0143dd049780-kube-api-access-ftvrv\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.449419 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.464990 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config" (OuterVolumeSpecName: "config") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.491302 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.502547 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.505711 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3a5b5b3a-6ec0-4454-9a80-0143dd049780" (UID: "3a5b5b3a-6ec0-4454-9a80-0143dd049780"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.533889 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.533933 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.533942 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.533950 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.533959 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a5b5b3a-6ec0-4454-9a80-0143dd049780-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.636206 4970 generic.go:334] "Generic (PLEG): container finished" podID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerID="377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d" exitCode=0 Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.636439 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.636504 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" event={"ID":"3a5b5b3a-6ec0-4454-9a80-0143dd049780","Type":"ContainerDied","Data":"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.636531 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66b66f7449-sf8l6" event={"ID":"3a5b5b3a-6ec0-4454-9a80-0143dd049780","Type":"ContainerDied","Data":"97990ddfacbfa144f3163ca751a3b371fcc62c7fb7b65ce9508cf57d8852f212"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.636546 4970 scope.go:117] "RemoveContainer" containerID="377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.648485 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" event={"ID":"df1d70bf-d612-4ef1-a113-6be419867e1d","Type":"ContainerStarted","Data":"6ce1c3062d1ac9c64dae5e20584224cfd194312baa7e160bf2b8eb9a23f80eba"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.657006 4970 generic.go:334] "Generic (PLEG): container finished" podID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerID="a076460ac0bd5ec78eec01f4ef65fe4f5acb9f6053f7cfaea8d0553ea7d8d65a" exitCode=0 Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.657647 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" event={"ID":"eb87ce82-23c4-4610-a607-c648b7c88b81","Type":"ContainerDied","Data":"a076460ac0bd5ec78eec01f4ef65fe4f5acb9f6053f7cfaea8d0553ea7d8d65a"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.675844 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85596cc87-lvc2b" event={"ID":"17b7d151-f71d-4f02-9d52-4c7ba81ab89e","Type":"ContainerStarted","Data":"6dfcb53317895aee3a3d068d4544c5b88ee5e83cf3086b7af8dfbb6daf7017da"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.677477 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-688cdc6cb6-fkrjb" podStartSLOduration=3.329127133 podStartE2EDuration="5.67745806s" podCreationTimestamp="2025-11-24 14:03:07 +0000 UTC" firstStartedPulling="2025-11-24 14:03:08.557839341 +0000 UTC m=+1003.845596634" lastFinishedPulling="2025-11-24 14:03:10.906170268 +0000 UTC m=+1006.193927561" observedRunningTime="2025-11-24 14:03:12.674623445 +0000 UTC m=+1007.962380738" watchObservedRunningTime="2025-11-24 14:03:12.67745806 +0000 UTC m=+1007.965215353" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.688897 4970 scope.go:117] "RemoveContainer" containerID="d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.692288 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerStarted","Data":"320e5ee7f641d4e70ff575a5a1f38b708b09c674151609021fa52e78f50ac920"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.700646 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.710274 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66b66f7449-sf8l6"] Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.725086 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerStarted","Data":"ab0e7b33253a97c28a38e45214fe61a4451b9340d1723954954ac1022788f0ad"} Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.781865 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-85596cc87-lvc2b" podStartSLOduration=3.475294008 podStartE2EDuration="5.781842277s" podCreationTimestamp="2025-11-24 14:03:07 +0000 UTC" firstStartedPulling="2025-11-24 14:03:08.548806302 +0000 UTC m=+1003.836563595" lastFinishedPulling="2025-11-24 14:03:10.855354571 +0000 UTC m=+1006.143111864" observedRunningTime="2025-11-24 14:03:12.765286968 +0000 UTC m=+1008.053044261" watchObservedRunningTime="2025-11-24 14:03:12.781842277 +0000 UTC m=+1008.069599570" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.784285 4970 scope.go:117] "RemoveContainer" containerID="377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d" Nov 24 14:03:12 crc kubenswrapper[4970]: E1124 14:03:12.788200 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d\": container with ID starting with 377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d not found: ID does not exist" containerID="377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.788247 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d"} err="failed to get container status \"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d\": rpc error: code = NotFound desc = could not find container \"377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d\": container with ID starting with 377e7abb422b651d16f0953f409d3301f0a49f82c69d11c0ed4106fc56c8c54d not found: ID does not exist" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.788289 4970 scope.go:117] "RemoveContainer" containerID="d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18" Nov 24 14:03:12 crc kubenswrapper[4970]: E1124 14:03:12.790688 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18\": container with ID starting with d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18 not found: ID does not exist" containerID="d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18" Nov 24 14:03:12 crc kubenswrapper[4970]: I1124 14:03:12.790718 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18"} err="failed to get container status \"d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18\": rpc error: code = NotFound desc = could not find container \"d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18\": container with ID starting with d410b181e41591275c5f833780fb558ec2e6ef043eb951dc78a63f3dfd747c18 not found: ID does not exist" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.482035 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" path="/var/lib/kubelet/pods/3a5b5b3a-6ec0-4454-9a80-0143dd049780/volumes" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.750222 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" event={"ID":"eb87ce82-23c4-4610-a607-c648b7c88b81","Type":"ContainerStarted","Data":"c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f"} Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.750638 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.761294 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerStarted","Data":"64ccdd89853f8523254826244733675046598c76ab002264fc4994c850d48c95"} Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.763609 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerStarted","Data":"294c443f968e953c1153ffb3126a66c7b80370eeda01347dcd2d6ad201936fcb"} Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.764016 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.768696 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerStarted","Data":"b60e14d93f47cdb012466987bf39667da80b69067d008c971b7a3ef4db1597d7"} Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.771373 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" podStartSLOduration=4.771355087 podStartE2EDuration="4.771355087s" podCreationTimestamp="2025-11-24 14:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:13.766028465 +0000 UTC m=+1009.053785758" watchObservedRunningTime="2025-11-24 14:03:13.771355087 +0000 UTC m=+1009.059112380" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.791186 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.791169102 podStartE2EDuration="4.791169102s" podCreationTimestamp="2025-11-24 14:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:13.788361787 +0000 UTC m=+1009.076119080" watchObservedRunningTime="2025-11-24 14:03:13.791169102 +0000 UTC m=+1009.078926395" Nov 24 14:03:13 crc kubenswrapper[4970]: I1124 14:03:13.952289 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.142231 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.330944 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.567323 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-756b7f5776-9gksb"] Nov 24 14:03:14 crc kubenswrapper[4970]: E1124 14:03:14.567778 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="init" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.567795 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="init" Nov 24 14:03:14 crc kubenswrapper[4970]: E1124 14:03:14.567828 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="dnsmasq-dns" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.567837 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="dnsmasq-dns" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.568084 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5b5b3a-6ec0-4454-9a80-0143dd049780" containerName="dnsmasq-dns" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.569243 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.572357 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.572720 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.582022 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756b7f5776-9gksb"] Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688463 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-internal-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688555 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data-custom\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688616 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-public-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688674 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83e2deca-2802-4021-9e47-b933ace28f4a-logs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688709 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lflbn\" (UniqueName: \"kubernetes.io/projected/83e2deca-2802-4021-9e47-b933ace28f4a-kube-api-access-lflbn\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688730 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-combined-ca-bundle\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.688802 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.779264 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerStarted","Data":"4e44ff873fbff05e81f28910559497edc3ff6ebbf8dae47449a1375062244bed"} Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.782017 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerStarted","Data":"3bfa59ab1ab3da913b632d7477d16cf65420809b8271e67474c9200693241ffc"} Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790051 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790101 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-internal-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790143 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data-custom\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790172 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-public-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790212 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83e2deca-2802-4021-9e47-b933ace28f4a-logs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790236 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lflbn\" (UniqueName: \"kubernetes.io/projected/83e2deca-2802-4021-9e47-b933ace28f4a-kube-api-access-lflbn\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790253 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-combined-ca-bundle\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.790682 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83e2deca-2802-4021-9e47-b933ace28f4a-logs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.796217 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-internal-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.797268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-public-tls-certs\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.800274 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data-custom\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.801091 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-config-data\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.812273 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83e2deca-2802-4021-9e47-b933ace28f4a-combined-ca-bundle\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.813701 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lflbn\" (UniqueName: \"kubernetes.io/projected/83e2deca-2802-4021-9e47-b933ace28f4a-kube-api-access-lflbn\") pod \"barbican-api-756b7f5776-9gksb\" (UID: \"83e2deca-2802-4021-9e47-b933ace28f4a\") " pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.827817 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.493714866 podStartE2EDuration="5.827792739s" podCreationTimestamp="2025-11-24 14:03:09 +0000 UTC" firstStartedPulling="2025-11-24 14:03:11.383041289 +0000 UTC m=+1006.670798582" lastFinishedPulling="2025-11-24 14:03:12.717119162 +0000 UTC m=+1008.004876455" observedRunningTime="2025-11-24 14:03:14.809643598 +0000 UTC m=+1010.097400901" watchObservedRunningTime="2025-11-24 14:03:14.827792739 +0000 UTC m=+1010.115550032" Nov 24 14:03:14 crc kubenswrapper[4970]: I1124 14:03:14.890531 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.030775 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.566433 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756b7f5776-9gksb"] Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.648158 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.792505 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerStarted","Data":"d4d7a2a412199f99b3bb9871a1e7c38164321d367a99cd85aef568645b701844"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.793850 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.803549 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756b7f5776-9gksb" event={"ID":"83e2deca-2802-4021-9e47-b933ace28f4a","Type":"ContainerStarted","Data":"bfdd42b0339242108a64747cc0a2c13770a2209a055dafcb90c5dca16745c7ea"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.803615 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756b7f5776-9gksb" event={"ID":"83e2deca-2802-4021-9e47-b933ace28f4a","Type":"ContainerStarted","Data":"0dfaa68543e67b0ad431aebe1b64f942ce32d95352b295aaee7e14c962dfd9a6"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807631 4970 generic.go:334] "Generic (PLEG): container finished" podID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerID="8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" exitCode=137 Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807658 4970 generic.go:334] "Generic (PLEG): container finished" podID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerID="5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" exitCode=137 Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807742 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d6c797b4f-hv87q" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807843 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerDied","Data":"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807883 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerDied","Data":"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807899 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d6c797b4f-hv87q" event={"ID":"f1c6bc70-16db-41fd-a10a-192f6a40d306","Type":"ContainerDied","Data":"9beaf39425d07fae76a5208913e684f62462f48f4c6c828964e2aa4304252fae"} Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.807921 4970 scope.go:117] "RemoveContainer" containerID="8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.808409 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api-log" containerID="cri-o://320e5ee7f641d4e70ff575a5a1f38b708b09c674151609021fa52e78f50ac920" gracePeriod=30 Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.808523 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api" containerID="cri-o://294c443f968e953c1153ffb3126a66c7b80370eeda01347dcd2d6ad201936fcb" gracePeriod=30 Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.812494 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs\") pod \"f1c6bc70-16db-41fd-a10a-192f6a40d306\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.812594 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key\") pod \"f1c6bc70-16db-41fd-a10a-192f6a40d306\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.812628 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data\") pod \"f1c6bc70-16db-41fd-a10a-192f6a40d306\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.812694 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts\") pod \"f1c6bc70-16db-41fd-a10a-192f6a40d306\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.813362 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t7js\" (UniqueName: \"kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js\") pod \"f1c6bc70-16db-41fd-a10a-192f6a40d306\" (UID: \"f1c6bc70-16db-41fd-a10a-192f6a40d306\") " Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.819325 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js" (OuterVolumeSpecName: "kube-api-access-9t7js") pod "f1c6bc70-16db-41fd-a10a-192f6a40d306" (UID: "f1c6bc70-16db-41fd-a10a-192f6a40d306"). InnerVolumeSpecName "kube-api-access-9t7js". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.820406 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f1c6bc70-16db-41fd-a10a-192f6a40d306" (UID: "f1c6bc70-16db-41fd-a10a-192f6a40d306"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.820571 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs" (OuterVolumeSpecName: "logs") pod "f1c6bc70-16db-41fd-a10a-192f6a40d306" (UID: "f1c6bc70-16db-41fd-a10a-192f6a40d306"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.834662 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.044546785 podStartE2EDuration="6.834633928s" podCreationTimestamp="2025-11-24 14:03:09 +0000 UTC" firstStartedPulling="2025-11-24 14:03:11.235708704 +0000 UTC m=+1006.523465997" lastFinishedPulling="2025-11-24 14:03:15.025795847 +0000 UTC m=+1010.313553140" observedRunningTime="2025-11-24 14:03:15.818839949 +0000 UTC m=+1011.106597252" watchObservedRunningTime="2025-11-24 14:03:15.834633928 +0000 UTC m=+1011.122391211" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.848373 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data" (OuterVolumeSpecName: "config-data") pod "f1c6bc70-16db-41fd-a10a-192f6a40d306" (UID: "f1c6bc70-16db-41fd-a10a-192f6a40d306"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.859753 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts" (OuterVolumeSpecName: "scripts") pod "f1c6bc70-16db-41fd-a10a-192f6a40d306" (UID: "f1c6bc70-16db-41fd-a10a-192f6a40d306"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.917719 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t7js\" (UniqueName: \"kubernetes.io/projected/f1c6bc70-16db-41fd-a10a-192f6a40d306-kube-api-access-9t7js\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.917756 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c6bc70-16db-41fd-a10a-192f6a40d306-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.917767 4970 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f1c6bc70-16db-41fd-a10a-192f6a40d306-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.917777 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:15 crc kubenswrapper[4970]: I1124 14:03:15.917788 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f1c6bc70-16db-41fd-a10a-192f6a40d306-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.144062 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.154842 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d6c797b4f-hv87q"] Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.215871 4970 scope.go:117] "RemoveContainer" containerID="5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.251869 4970 scope.go:117] "RemoveContainer" containerID="8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" Nov 24 14:03:16 crc kubenswrapper[4970]: E1124 14:03:16.252826 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da\": container with ID starting with 8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da not found: ID does not exist" containerID="8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.252872 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da"} err="failed to get container status \"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da\": rpc error: code = NotFound desc = could not find container \"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da\": container with ID starting with 8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da not found: ID does not exist" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.252904 4970 scope.go:117] "RemoveContainer" containerID="5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" Nov 24 14:03:16 crc kubenswrapper[4970]: E1124 14:03:16.253193 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8\": container with ID starting with 5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8 not found: ID does not exist" containerID="5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.253224 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8"} err="failed to get container status \"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8\": rpc error: code = NotFound desc = could not find container \"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8\": container with ID starting with 5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8 not found: ID does not exist" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.253249 4970 scope.go:117] "RemoveContainer" containerID="8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.253486 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da"} err="failed to get container status \"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da\": rpc error: code = NotFound desc = could not find container \"8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da\": container with ID starting with 8009c9bbfab3323e75c25a27eeaa979c27e75b0f171cfc55d09858bc2849a6da not found: ID does not exist" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.253506 4970 scope.go:117] "RemoveContainer" containerID="5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.253792 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8"} err="failed to get container status \"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8\": rpc error: code = NotFound desc = could not find container \"5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8\": container with ID starting with 5ca4f3c9c2f1c27d13b77663db5c49a0533a66e2824c6fb3244ff8b8cc6d3db8 not found: ID does not exist" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.762128 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.835281 4970 generic.go:334] "Generic (PLEG): container finished" podID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerID="294c443f968e953c1153ffb3126a66c7b80370eeda01347dcd2d6ad201936fcb" exitCode=0 Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.835307 4970 generic.go:334] "Generic (PLEG): container finished" podID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerID="320e5ee7f641d4e70ff575a5a1f38b708b09c674151609021fa52e78f50ac920" exitCode=143 Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.835332 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerDied","Data":"294c443f968e953c1153ffb3126a66c7b80370eeda01347dcd2d6ad201936fcb"} Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.835417 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerDied","Data":"320e5ee7f641d4e70ff575a5a1f38b708b09c674151609021fa52e78f50ac920"} Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.850207 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756b7f5776-9gksb" event={"ID":"83e2deca-2802-4021-9e47-b933ace28f4a","Type":"ContainerStarted","Data":"6b3610622080fdec0e05c8c9d1312b6d0a61dffa5ae72123d3c970e21a20fe4c"} Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.850413 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.850443 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:16 crc kubenswrapper[4970]: I1124 14:03:16.878289 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-756b7f5776-9gksb" podStartSLOduration=2.878271592 podStartE2EDuration="2.878271592s" podCreationTimestamp="2025-11-24 14:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:16.875130288 +0000 UTC m=+1012.162887591" watchObservedRunningTime="2025-11-24 14:03:16.878271592 +0000 UTC m=+1012.166028885" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.225184 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5bb8f9fccd-2m5xk" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.225528 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.288887 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.289135 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon-log" containerID="cri-o://ca522efedbf9731e67c7b3e65f8e485ba5ff868b422ba102b15ddfee9e01717d" gracePeriod=30 Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.289603 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" containerID="cri-o://6f5a1170f608d84dfefcb3970f5b308afb8b70329afa1e99dcc81638e534855d" gracePeriod=30 Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344321 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344381 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p2cz\" (UniqueName: \"kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344447 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344511 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344541 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344610 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344655 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data\") pod \"f9dbf176-65e5-4236-97f7-ce326889c4d2\" (UID: \"f9dbf176-65e5-4236-97f7-ce326889c4d2\") " Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.344805 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.345142 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs" (OuterVolumeSpecName: "logs") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.345395 4970 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9dbf176-65e5-4236-97f7-ce326889c4d2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.345408 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9dbf176-65e5-4236-97f7-ce326889c4d2-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.351516 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz" (OuterVolumeSpecName: "kube-api-access-7p2cz") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "kube-api-access-7p2cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.353003 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.353744 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts" (OuterVolumeSpecName: "scripts") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.382760 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.443041 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data" (OuterVolumeSpecName: "config-data") pod "f9dbf176-65e5-4236-97f7-ce326889c4d2" (UID: "f9dbf176-65e5-4236-97f7-ce326889c4d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.447078 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.447100 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p2cz\" (UniqueName: \"kubernetes.io/projected/f9dbf176-65e5-4236-97f7-ce326889c4d2-kube-api-access-7p2cz\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.447113 4970 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.447122 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.447130 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9dbf176-65e5-4236-97f7-ce326889c4d2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.482175 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" path="/var/lib/kubelet/pods/f1c6bc70-16db-41fd-a10a-192f6a40d306/volumes" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.629354 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.873942 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.874557 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f9dbf176-65e5-4236-97f7-ce326889c4d2","Type":"ContainerDied","Data":"6b32fcb5e9c8880188200d97b154c76c75b481f3159e447f2a34bc26c224f1b5"} Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.874618 4970 scope.go:117] "RemoveContainer" containerID="294c443f968e953c1153ffb3126a66c7b80370eeda01347dcd2d6ad201936fcb" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.917275 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.921704 4970 scope.go:117] "RemoveContainer" containerID="320e5ee7f641d4e70ff575a5a1f38b708b09c674151609021fa52e78f50ac920" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.929514 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.948150 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:17 crc kubenswrapper[4970]: E1124 14:03:17.948570 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api-log" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949165 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api-log" Nov 24 14:03:17 crc kubenswrapper[4970]: E1124 14:03:17.949198 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949210 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api" Nov 24 14:03:17 crc kubenswrapper[4970]: E1124 14:03:17.949227 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon-log" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949238 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon-log" Nov 24 14:03:17 crc kubenswrapper[4970]: E1124 14:03:17.949258 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949266 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949454 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api-log" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949472 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" containerName="cinder-api" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949488 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.949497 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c6bc70-16db-41fd-a10a-192f6a40d306" containerName="horizon-log" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.952206 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.957005 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.957995 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.958158 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 14:03:17 crc kubenswrapper[4970]: I1124 14:03:17.958285 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057351 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data-custom\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057481 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057520 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c4a13-7335-4e6f-b424-90a6a003a894-logs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057560 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057664 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6c4a13-7335-4e6f-b424-90a6a003a894-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057702 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fmlh\" (UniqueName: \"kubernetes.io/projected/9c6c4a13-7335-4e6f-b424-90a6a003a894-kube-api-access-4fmlh\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057759 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.057835 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-scripts\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.058029 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159251 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159320 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data-custom\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159363 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159383 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c4a13-7335-4e6f-b424-90a6a003a894-logs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159401 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159422 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6c4a13-7335-4e6f-b424-90a6a003a894-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159440 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fmlh\" (UniqueName: \"kubernetes.io/projected/9c6c4a13-7335-4e6f-b424-90a6a003a894-kube-api-access-4fmlh\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159464 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.159493 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-scripts\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.160108 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9c6c4a13-7335-4e6f-b424-90a6a003a894-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.160644 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c4a13-7335-4e6f-b424-90a6a003a894-logs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.163640 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.164026 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-scripts\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.164848 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.166153 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.166951 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.175278 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c4a13-7335-4e6f-b424-90a6a003a894-config-data-custom\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.177976 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fmlh\" (UniqueName: \"kubernetes.io/projected/9c6c4a13-7335-4e6f-b424-90a6a003a894-kube-api-access-4fmlh\") pod \"cinder-api-0\" (UID: \"9c6c4a13-7335-4e6f-b424-90a6a003a894\") " pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.289520 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 14:03:18 crc kubenswrapper[4970]: W1124 14:03:18.707843 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c6c4a13_7335_4e6f_b424_90a6a003a894.slice/crio-79f6d11d26c8caa57f5243d2b3de1aeb3269255b8e942157b398e5a3c268970b WatchSource:0}: Error finding container 79f6d11d26c8caa57f5243d2b3de1aeb3269255b8e942157b398e5a3c268970b: Status 404 returned error can't find the container with id 79f6d11d26c8caa57f5243d2b3de1aeb3269255b8e942157b398e5a3c268970b Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.711913 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 14:03:18 crc kubenswrapper[4970]: I1124 14:03:18.900468 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9c6c4a13-7335-4e6f-b424-90a6a003a894","Type":"ContainerStarted","Data":"79f6d11d26c8caa57f5243d2b3de1aeb3269255b8e942157b398e5a3c268970b"} Nov 24 14:03:19 crc kubenswrapper[4970]: I1124 14:03:19.479895 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9dbf176-65e5-4236-97f7-ce326889c4d2" path="/var/lib/kubelet/pods/f9dbf176-65e5-4236-97f7-ce326889c4d2/volumes" Nov 24 14:03:19 crc kubenswrapper[4970]: I1124 14:03:19.880027 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:19 crc kubenswrapper[4970]: I1124 14:03:19.930378 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9c6c4a13-7335-4e6f-b424-90a6a003a894","Type":"ContainerStarted","Data":"814ba52019ab552d3fe1c73a895e46eae7ea6a60fa36e24f9dc421229d86d0db"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.071720 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6464bbf59c-l44tb" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.086519 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.164803 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.165391 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66b7fdd9cd-76pz9" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-api" containerID="cri-o://25d030299d57e4ef5ade75a2cba774943a4477dcb0e7d4d461d152708684cf0f" gracePeriod=30 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.166101 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66b7fdd9cd-76pz9" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-httpd" containerID="cri-o://c48df9554a6eac5c7de114da9d4e9d3d33153f4f11c4a691480a39cf0cd01b1d" gracePeriod=30 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.217735 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.281297 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.281557 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="dnsmasq-dns" containerID="cri-o://341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e" gracePeriod=10 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.512451 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.548064 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.564789 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.928204 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.966475 4970 generic.go:334] "Generic (PLEG): container finished" podID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerID="341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e" exitCode=0 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.966559 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" event={"ID":"1bcbbd0d-82de-4751-b0ca-aae14c1f1337","Type":"ContainerDied","Data":"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.966606 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" event={"ID":"1bcbbd0d-82de-4751-b0ca-aae14c1f1337","Type":"ContainerDied","Data":"d75375e48d12c89e63768fbf221c3a61827566b4d1bf092914c37569cb52e735"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.966627 4970 scope.go:117] "RemoveContainer" containerID="341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.966561 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b95cfcf9c-7t8hv" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.969233 4970 generic.go:334] "Generic (PLEG): container finished" podID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerID="6f5a1170f608d84dfefcb3970f5b308afb8b70329afa1e99dcc81638e534855d" exitCode=0 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.969259 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerDied","Data":"6f5a1170f608d84dfefcb3970f5b308afb8b70329afa1e99dcc81638e534855d"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.971063 4970 generic.go:334] "Generic (PLEG): container finished" podID="68622514-f76c-41a6-8262-99155e3647ed" containerID="c48df9554a6eac5c7de114da9d4e9d3d33153f4f11c4a691480a39cf0cd01b1d" exitCode=0 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.971104 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerDied","Data":"c48df9554a6eac5c7de114da9d4e9d3d33153f4f11c4a691480a39cf0cd01b1d"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.984309 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="cinder-scheduler" containerID="cri-o://64ccdd89853f8523254826244733675046598c76ab002264fc4994c850d48c95" gracePeriod=30 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.986597 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9c6c4a13-7335-4e6f-b424-90a6a003a894","Type":"ContainerStarted","Data":"ab49a365f962e9fa639dfb6a992fb6ece627b319e3f2ed48eb0b0a36f0a0246e"} Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.986649 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.986981 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="probe" containerID="cri-o://4e44ff873fbff05e81f28910559497edc3ff6ebbf8dae47449a1375062244bed" gracePeriod=30 Nov 24 14:03:20 crc kubenswrapper[4970]: I1124 14:03:20.996781 4970 scope.go:117] "RemoveContainer" containerID="569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.013909 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.013896715 podStartE2EDuration="4.013896715s" podCreationTimestamp="2025-11-24 14:03:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:21.012257502 +0000 UTC m=+1016.300014795" watchObservedRunningTime="2025-11-24 14:03:21.013896715 +0000 UTC m=+1016.301654008" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030240 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030310 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030378 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t7wj\" (UniqueName: \"kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030483 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030517 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.030573 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb\") pod \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\" (UID: \"1bcbbd0d-82de-4751-b0ca-aae14c1f1337\") " Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.047851 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj" (OuterVolumeSpecName: "kube-api-access-9t7wj") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "kube-api-access-9t7wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.073461 4970 scope.go:117] "RemoveContainer" containerID="341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e" Nov 24 14:03:21 crc kubenswrapper[4970]: E1124 14:03:21.073829 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e\": container with ID starting with 341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e not found: ID does not exist" containerID="341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.073851 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e"} err="failed to get container status \"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e\": rpc error: code = NotFound desc = could not find container \"341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e\": container with ID starting with 341ccea44d01e420838fa81e2509b83da917458b34c6f5cfb6076ddf97a3952e not found: ID does not exist" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.073869 4970 scope.go:117] "RemoveContainer" containerID="569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af" Nov 24 14:03:21 crc kubenswrapper[4970]: E1124 14:03:21.074056 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af\": container with ID starting with 569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af not found: ID does not exist" containerID="569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.074070 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af"} err="failed to get container status \"569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af\": rpc error: code = NotFound desc = could not find container \"569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af\": container with ID starting with 569b35d614d7403c97351a99fc19ac01f535e4edf1543a32787d4f36fa25a3af not found: ID does not exist" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.090374 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.117452 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.120167 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.120718 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.131348 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config" (OuterVolumeSpecName: "config") pod "1bcbbd0d-82de-4751-b0ca-aae14c1f1337" (UID: "1bcbbd0d-82de-4751-b0ca-aae14c1f1337"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132690 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132710 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t7wj\" (UniqueName: \"kubernetes.io/projected/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-kube-api-access-9t7wj\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132823 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132833 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132842 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.132851 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1bcbbd0d-82de-4751-b0ca-aae14c1f1337-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.299922 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.308320 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b95cfcf9c-7t8hv"] Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.494073 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" path="/var/lib/kubelet/pods/1bcbbd0d-82de-4751-b0ca-aae14c1f1337/volumes" Nov 24 14:03:21 crc kubenswrapper[4970]: I1124 14:03:21.696283 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.007151 4970 generic.go:334] "Generic (PLEG): container finished" podID="70c613a6-e813-420a-9fc3-2b993ac73787" containerID="4e44ff873fbff05e81f28910559497edc3ff6ebbf8dae47449a1375062244bed" exitCode=0 Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.007186 4970 generic.go:334] "Generic (PLEG): container finished" podID="70c613a6-e813-420a-9fc3-2b993ac73787" containerID="64ccdd89853f8523254826244733675046598c76ab002264fc4994c850d48c95" exitCode=0 Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.007269 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerDied","Data":"4e44ff873fbff05e81f28910559497edc3ff6ebbf8dae47449a1375062244bed"} Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.007294 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerDied","Data":"64ccdd89853f8523254826244733675046598c76ab002264fc4994c850d48c95"} Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.395079 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463486 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463542 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463601 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463623 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463668 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s565z\" (UniqueName: \"kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463716 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle\") pod \"70c613a6-e813-420a-9fc3-2b993ac73787\" (UID: \"70c613a6-e813-420a-9fc3-2b993ac73787\") " Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.463769 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.464076 4970 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70c613a6-e813-420a-9fc3-2b993ac73787-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.471176 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts" (OuterVolumeSpecName: "scripts") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.473158 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z" (OuterVolumeSpecName: "kube-api-access-s565z") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "kube-api-access-s565z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.493791 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.566205 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.566243 4970 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.566257 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s565z\" (UniqueName: \"kubernetes.io/projected/70c613a6-e813-420a-9fc3-2b993ac73787-kube-api-access-s565z\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.586255 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data" (OuterVolumeSpecName: "config-data") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.588723 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70c613a6-e813-420a-9fc3-2b993ac73787" (UID: "70c613a6-e813-420a-9fc3-2b993ac73787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.668233 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:22 crc kubenswrapper[4970]: I1124 14:03:22.668279 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c613a6-e813-420a-9fc3-2b993ac73787-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.025217 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70c613a6-e813-420a-9fc3-2b993ac73787","Type":"ContainerDied","Data":"31a83a94e797fe68fee9a338ad091cb4068cd8b49fe6398ddf9986d7160cecce"} Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.025265 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.025273 4970 scope.go:117] "RemoveContainer" containerID="4e44ff873fbff05e81f28910559497edc3ff6ebbf8dae47449a1375062244bed" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.071883 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.086663 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.112644 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:23 crc kubenswrapper[4970]: E1124 14:03:23.113092 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="dnsmasq-dns" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113110 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="dnsmasq-dns" Nov 24 14:03:23 crc kubenswrapper[4970]: E1124 14:03:23.113139 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="init" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113169 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="init" Nov 24 14:03:23 crc kubenswrapper[4970]: E1124 14:03:23.113177 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="probe" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113184 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="probe" Nov 24 14:03:23 crc kubenswrapper[4970]: E1124 14:03:23.113202 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="cinder-scheduler" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113209 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="cinder-scheduler" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113414 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bcbbd0d-82de-4751-b0ca-aae14c1f1337" containerName="dnsmasq-dns" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113441 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="cinder-scheduler" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.113454 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" containerName="probe" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.114375 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.124915 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.131996 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177370 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8304c7a-67d5-4d0f-afce-231f09bf48ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177520 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177591 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177634 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177689 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swr52\" (UniqueName: \"kubernetes.io/projected/c8304c7a-67d5-4d0f-afce-231f09bf48ea-kube-api-access-swr52\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.177723 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.228296 4970 scope.go:117] "RemoveContainer" containerID="64ccdd89853f8523254826244733675046598c76ab002264fc4994c850d48c95" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279170 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swr52\" (UniqueName: \"kubernetes.io/projected/c8304c7a-67d5-4d0f-afce-231f09bf48ea-kube-api-access-swr52\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279234 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279277 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8304c7a-67d5-4d0f-afce-231f09bf48ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279378 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279411 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c8304c7a-67d5-4d0f-afce-231f09bf48ea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279436 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.279472 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.284217 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.284224 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-scripts\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.284374 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.284600 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8304c7a-67d5-4d0f-afce-231f09bf48ea-config-data\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.305313 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swr52\" (UniqueName: \"kubernetes.io/projected/c8304c7a-67d5-4d0f-afce-231f09bf48ea-kube-api-access-swr52\") pod \"cinder-scheduler-0\" (UID: \"c8304c7a-67d5-4d0f-afce-231f09bf48ea\") " pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.456104 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.486352 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c613a6-e813-420a-9fc3-2b993ac73787" path="/var/lib/kubelet/pods/70c613a6-e813-420a-9fc3-2b993ac73787/volumes" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.865596 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-8c975fcc4-8cmzb" Nov 24 14:03:23 crc kubenswrapper[4970]: I1124 14:03:23.939693 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 14:03:23 crc kubenswrapper[4970]: W1124 14:03:23.950672 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8304c7a_67d5_4d0f_afce_231f09bf48ea.slice/crio-1b01c5f26ddc930d634b2522100b35d74bb5eb54ead78fa7b250a6dfbf1d37a2 WatchSource:0}: Error finding container 1b01c5f26ddc930d634b2522100b35d74bb5eb54ead78fa7b250a6dfbf1d37a2: Status 404 returned error can't find the container with id 1b01c5f26ddc930d634b2522100b35d74bb5eb54ead78fa7b250a6dfbf1d37a2 Nov 24 14:03:24 crc kubenswrapper[4970]: I1124 14:03:24.058606 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c8304c7a-67d5-4d0f-afce-231f09bf48ea","Type":"ContainerStarted","Data":"1b01c5f26ddc930d634b2522100b35d74bb5eb54ead78fa7b250a6dfbf1d37a2"} Nov 24 14:03:25 crc kubenswrapper[4970]: I1124 14:03:25.082125 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c8304c7a-67d5-4d0f-afce-231f09bf48ea","Type":"ContainerStarted","Data":"a78acaa484622ce49b003eff669c05cfa41b5172fbdcc4ccd55d952800952689"} Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.094145 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c8304c7a-67d5-4d0f-afce-231f09bf48ea","Type":"ContainerStarted","Data":"59a59d9dc9cb7530f280f3bcb082d76bd6717d1f67a6b2b09a72c7eee59d8635"} Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.168142 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.168125579 podStartE2EDuration="3.168125579s" podCreationTimestamp="2025-11-24 14:03:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:26.117807626 +0000 UTC m=+1021.405564909" watchObservedRunningTime="2025-11-24 14:03:26.168125579 +0000 UTC m=+1021.455882872" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.172943 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.174020 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.178439 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-q8xc5" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.178478 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.178730 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.188063 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.248142 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.248231 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfcvh\" (UniqueName: \"kubernetes.io/projected/0c4f17a4-9a8d-4824-8370-6aea04d166ae-kube-api-access-dfcvh\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.248334 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.248424 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.350031 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.350129 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.350192 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.350220 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfcvh\" (UniqueName: \"kubernetes.io/projected/0c4f17a4-9a8d-4824-8370-6aea04d166ae-kube-api-access-dfcvh\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.351066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.369299 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.371702 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0c4f17a4-9a8d-4824-8370-6aea04d166ae-openstack-config-secret\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.376772 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfcvh\" (UniqueName: \"kubernetes.io/projected/0c4f17a4-9a8d-4824-8370-6aea04d166ae-kube-api-access-dfcvh\") pod \"openstackclient\" (UID: \"0c4f17a4-9a8d-4824-8370-6aea04d166ae\") " pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.509757 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.708220 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.784098 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756b7f5776-9gksb" Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.855549 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.856306 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api" containerID="cri-o://159b202c0d4d83b56216172cee771df85c777db30392ab8e1cafb700255dab60" gracePeriod=30 Nov 24 14:03:26 crc kubenswrapper[4970]: I1124 14:03:26.855914 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api-log" containerID="cri-o://532b30d02f415d46f0a49cf7392567509233fd9aaec005abeab067ebdf632840" gracePeriod=30 Nov 24 14:03:27 crc kubenswrapper[4970]: W1124 14:03:27.063018 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c4f17a4_9a8d_4824_8370_6aea04d166ae.slice/crio-f253cb7564f29c3a842e02aaa8039fc695d3084b5362e7e4b1a5a09192318a03 WatchSource:0}: Error finding container f253cb7564f29c3a842e02aaa8039fc695d3084b5362e7e4b1a5a09192318a03: Status 404 returned error can't find the container with id f253cb7564f29c3a842e02aaa8039fc695d3084b5362e7e4b1a5a09192318a03 Nov 24 14:03:27 crc kubenswrapper[4970]: I1124 14:03:27.064968 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 14:03:27 crc kubenswrapper[4970]: I1124 14:03:27.104013 4970 generic.go:334] "Generic (PLEG): container finished" podID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerID="532b30d02f415d46f0a49cf7392567509233fd9aaec005abeab067ebdf632840" exitCode=143 Nov 24 14:03:27 crc kubenswrapper[4970]: I1124 14:03:27.104087 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerDied","Data":"532b30d02f415d46f0a49cf7392567509233fd9aaec005abeab067ebdf632840"} Nov 24 14:03:27 crc kubenswrapper[4970]: I1124 14:03:27.108668 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0c4f17a4-9a8d-4824-8370-6aea04d166ae","Type":"ContainerStarted","Data":"f253cb7564f29c3a842e02aaa8039fc695d3084b5362e7e4b1a5a09192318a03"} Nov 24 14:03:28 crc kubenswrapper[4970]: I1124 14:03:28.456534 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.019848 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:37132->10.217.0.164:9311: read: connection reset by peer" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.019848 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:37128->10.217.0.164:9311: read: connection reset by peer" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.147111 4970 generic.go:334] "Generic (PLEG): container finished" podID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerID="159b202c0d4d83b56216172cee771df85c777db30392ab8e1cafb700255dab60" exitCode=0 Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.147163 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerDied","Data":"159b202c0d4d83b56216172cee771df85c777db30392ab8e1cafb700255dab60"} Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.448916 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.519376 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.680480 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle\") pod \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.680542 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom\") pod \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.680656 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mxck\" (UniqueName: \"kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck\") pod \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.680712 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs\") pod \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.680805 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data\") pod \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\" (UID: \"2ab96cbb-2ec7-4e63-8772-6be2ec119b79\") " Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.681288 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs" (OuterVolumeSpecName: "logs") pod "2ab96cbb-2ec7-4e63-8772-6be2ec119b79" (UID: "2ab96cbb-2ec7-4e63-8772-6be2ec119b79"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.687764 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2ab96cbb-2ec7-4e63-8772-6be2ec119b79" (UID: "2ab96cbb-2ec7-4e63-8772-6be2ec119b79"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.687837 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck" (OuterVolumeSpecName: "kube-api-access-2mxck") pod "2ab96cbb-2ec7-4e63-8772-6be2ec119b79" (UID: "2ab96cbb-2ec7-4e63-8772-6be2ec119b79"). InnerVolumeSpecName "kube-api-access-2mxck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.719628 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ab96cbb-2ec7-4e63-8772-6be2ec119b79" (UID: "2ab96cbb-2ec7-4e63-8772-6be2ec119b79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.754757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data" (OuterVolumeSpecName: "config-data") pod "2ab96cbb-2ec7-4e63-8772-6be2ec119b79" (UID: "2ab96cbb-2ec7-4e63-8772-6be2ec119b79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.783571 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mxck\" (UniqueName: \"kubernetes.io/projected/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-kube-api-access-2mxck\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.783626 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.783639 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.783651 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:30 crc kubenswrapper[4970]: I1124 14:03:30.783662 4970 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ab96cbb-2ec7-4e63-8772-6be2ec119b79-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.143013 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-76bd896ccf-szjxt"] Nov 24 14:03:31 crc kubenswrapper[4970]: E1124 14:03:31.143472 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api-log" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.143485 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api-log" Nov 24 14:03:31 crc kubenswrapper[4970]: E1124 14:03:31.143513 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.143520 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.143728 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api-log" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.143749 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" containerName="barbican-api" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.145530 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.159168 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-76bd896ccf-szjxt"] Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.196011 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.196754 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.196888 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.230372 4970 generic.go:334] "Generic (PLEG): container finished" podID="68622514-f76c-41a6-8262-99155e3647ed" containerID="25d030299d57e4ef5ade75a2cba774943a4477dcb0e7d4d461d152708684cf0f" exitCode=0 Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.230437 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerDied","Data":"25d030299d57e4ef5ade75a2cba774943a4477dcb0e7d4d461d152708684cf0f"} Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.230462 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66b7fdd9cd-76pz9" event={"ID":"68622514-f76c-41a6-8262-99155e3647ed","Type":"ContainerDied","Data":"e32bf6d9fdbf0d20ef614eae6800597e898710ee4f305b4b9ead0035d48e7912"} Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.230472 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e32bf6d9fdbf0d20ef614eae6800597e898710ee4f305b4b9ead0035d48e7912" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.231251 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.234949 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" event={"ID":"2ab96cbb-2ec7-4e63-8772-6be2ec119b79","Type":"ContainerDied","Data":"52d415d8336b1ffbe87854e337814c68bee3de0d99c00311c047f093e5438067"} Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.235019 4970 scope.go:117] "RemoveContainer" containerID="159b202c0d4d83b56216172cee771df85c777db30392ab8e1cafb700255dab60" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.235099 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-64b6b9c8c6-hdl7c" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.284563 4970 scope.go:117] "RemoveContainer" containerID="532b30d02f415d46f0a49cf7392567509233fd9aaec005abeab067ebdf632840" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293300 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qkvn\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-kube-api-access-2qkvn\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293391 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-combined-ca-bundle\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293435 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-config-data\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293462 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-etc-swift\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293483 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-run-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293512 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-public-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293564 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-internal-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.293601 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-log-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.313292 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.323636 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-64b6b9c8c6-hdl7c"] Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.394520 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config\") pod \"68622514-f76c-41a6-8262-99155e3647ed\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.394618 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config\") pod \"68622514-f76c-41a6-8262-99155e3647ed\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.394796 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs\") pod \"68622514-f76c-41a6-8262-99155e3647ed\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.394839 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle\") pod \"68622514-f76c-41a6-8262-99155e3647ed\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.394905 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bjhd\" (UniqueName: \"kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd\") pod \"68622514-f76c-41a6-8262-99155e3647ed\" (UID: \"68622514-f76c-41a6-8262-99155e3647ed\") " Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395190 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-combined-ca-bundle\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395553 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-config-data\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395607 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-etc-swift\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395634 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-run-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395666 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-public-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395731 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-internal-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395768 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-log-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.395865 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qkvn\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-kube-api-access-2qkvn\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.402754 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "68622514-f76c-41a6-8262-99155e3647ed" (UID: "68622514-f76c-41a6-8262-99155e3647ed"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.404030 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-etc-swift\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.404605 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-log-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.404912 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32811b6b-86a2-4762-94fd-2eef9eedad76-run-httpd\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.411586 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-public-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.411992 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-combined-ca-bundle\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.412804 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd" (OuterVolumeSpecName: "kube-api-access-6bjhd") pod "68622514-f76c-41a6-8262-99155e3647ed" (UID: "68622514-f76c-41a6-8262-99155e3647ed"). InnerVolumeSpecName "kube-api-access-6bjhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.417240 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-config-data\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.421597 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qkvn\" (UniqueName: \"kubernetes.io/projected/32811b6b-86a2-4762-94fd-2eef9eedad76-kube-api-access-2qkvn\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.421762 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32811b6b-86a2-4762-94fd-2eef9eedad76-internal-tls-certs\") pod \"swift-proxy-76bd896ccf-szjxt\" (UID: \"32811b6b-86a2-4762-94fd-2eef9eedad76\") " pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.492167 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ab96cbb-2ec7-4e63-8772-6be2ec119b79" path="/var/lib/kubelet/pods/2ab96cbb-2ec7-4e63-8772-6be2ec119b79/volumes" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.498864 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.498897 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bjhd\" (UniqueName: \"kubernetes.io/projected/68622514-f76c-41a6-8262-99155e3647ed-kube-api-access-6bjhd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.513803 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config" (OuterVolumeSpecName: "config") pod "68622514-f76c-41a6-8262-99155e3647ed" (UID: "68622514-f76c-41a6-8262-99155e3647ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.523719 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68622514-f76c-41a6-8262-99155e3647ed" (UID: "68622514-f76c-41a6-8262-99155e3647ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.572712 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "68622514-f76c-41a6-8262-99155e3647ed" (UID: "68622514-f76c-41a6-8262-99155e3647ed"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.598014 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.600596 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.600621 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.600633 4970 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/68622514-f76c-41a6-8262-99155e3647ed-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.693851 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.694169 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-central-agent" containerID="cri-o://ab0e7b33253a97c28a38e45214fe61a4451b9340d1723954954ac1022788f0ad" gracePeriod=30 Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.695812 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" containerID="cri-o://d4d7a2a412199f99b3bb9871a1e7c38164321d367a99cd85aef568645b701844" gracePeriod=30 Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.695820 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-notification-agent" containerID="cri-o://b60e14d93f47cdb012466987bf39667da80b69067d008c971b7a3ef4db1597d7" gracePeriod=30 Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.695930 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="sg-core" containerID="cri-o://3bfa59ab1ab3da913b632d7477d16cf65420809b8271e67474c9200693241ffc" gracePeriod=30 Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.698168 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 14:03:31 crc kubenswrapper[4970]: I1124 14:03:31.712967 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": EOF" Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.210236 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-76bd896ccf-szjxt"] Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.251756 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76bd896ccf-szjxt" event={"ID":"32811b6b-86a2-4762-94fd-2eef9eedad76","Type":"ContainerStarted","Data":"2e02fa8348668d089630b2a5eac2294488bb796af23dc3d8774bed382c61f7cf"} Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280036 4970 generic.go:334] "Generic (PLEG): container finished" podID="9cfd8b77-9869-4672-865e-e39e8a752567" containerID="d4d7a2a412199f99b3bb9871a1e7c38164321d367a99cd85aef568645b701844" exitCode=0 Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280360 4970 generic.go:334] "Generic (PLEG): container finished" podID="9cfd8b77-9869-4672-865e-e39e8a752567" containerID="3bfa59ab1ab3da913b632d7477d16cf65420809b8271e67474c9200693241ffc" exitCode=2 Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280368 4970 generic.go:334] "Generic (PLEG): container finished" podID="9cfd8b77-9869-4672-865e-e39e8a752567" containerID="ab0e7b33253a97c28a38e45214fe61a4451b9340d1723954954ac1022788f0ad" exitCode=0 Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280469 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66b7fdd9cd-76pz9" Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280787 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerDied","Data":"d4d7a2a412199f99b3bb9871a1e7c38164321d367a99cd85aef568645b701844"} Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280862 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerDied","Data":"3bfa59ab1ab3da913b632d7477d16cf65420809b8271e67474c9200693241ffc"} Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.280877 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerDied","Data":"ab0e7b33253a97c28a38e45214fe61a4451b9340d1723954954ac1022788f0ad"} Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.351438 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:03:32 crc kubenswrapper[4970]: I1124 14:03:32.362161 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66b7fdd9cd-76pz9"] Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.292062 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76bd896ccf-szjxt" event={"ID":"32811b6b-86a2-4762-94fd-2eef9eedad76","Type":"ContainerStarted","Data":"38c78990b859b26d3138abc69215dd46318363a97165162ef2472e8372c68554"} Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.292304 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-76bd896ccf-szjxt" event={"ID":"32811b6b-86a2-4762-94fd-2eef9eedad76","Type":"ContainerStarted","Data":"63d0c4cd0deccf0f56ef10c95a96665ea1aaf933f2976ee58cc2fbd87869fc4b"} Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.292647 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.292666 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.319108 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-76bd896ccf-szjxt" podStartSLOduration=2.319093183 podStartE2EDuration="2.319093183s" podCreationTimestamp="2025-11-24 14:03:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:33.311118982 +0000 UTC m=+1028.598876275" watchObservedRunningTime="2025-11-24 14:03:33.319093183 +0000 UTC m=+1028.606850476" Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.482770 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68622514-f76c-41a6-8262-99155e3647ed" path="/var/lib/kubelet/pods/68622514-f76c-41a6-8262-99155e3647ed/volumes" Nov 24 14:03:33 crc kubenswrapper[4970]: I1124 14:03:33.735722 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.336036 4970 generic.go:334] "Generic (PLEG): container finished" podID="9cfd8b77-9869-4672-865e-e39e8a752567" containerID="b60e14d93f47cdb012466987bf39667da80b69067d008c971b7a3ef4db1597d7" exitCode=0 Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.336121 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerDied","Data":"b60e14d93f47cdb012466987bf39667da80b69067d008c971b7a3ef4db1597d7"} Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.367766 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hbxpk"] Nov 24 14:03:36 crc kubenswrapper[4970]: E1124 14:03:36.368228 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-httpd" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.368251 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-httpd" Nov 24 14:03:36 crc kubenswrapper[4970]: E1124 14:03:36.368273 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-api" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.368282 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-api" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.368530 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-api" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.368556 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="68622514-f76c-41a6-8262-99155e3647ed" containerName="neutron-httpd" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.369334 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.378801 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hbxpk"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.502627 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-tfqlt"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.503811 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.511081 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-84e1-account-create-f27tz"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.512405 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.516469 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.518634 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tfqlt"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.532071 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.532212 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87xl\" (UniqueName: \"kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.532622 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-84e1-account-create-f27tz"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.634712 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.634768 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vhlr\" (UniqueName: \"kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.634796 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57g6r\" (UniqueName: \"kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.634814 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.635813 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87xl\" (UniqueName: \"kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.635953 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.639185 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.662555 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b5b2b"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.664074 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.664807 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87xl\" (UniqueName: \"kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl\") pod \"nova-api-db-create-hbxpk\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.669987 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9849-account-create-nkhkx"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.673598 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.677624 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.691142 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b5b2b"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.699659 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9849-account-create-nkhkx"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.741507 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vhlr\" (UniqueName: \"kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.741646 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57g6r\" (UniqueName: \"kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.741672 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.742155 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.742969 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.744402 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.765071 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57g6r\" (UniqueName: \"kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r\") pod \"nova-api-84e1-account-create-f27tz\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.766096 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.790002 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vhlr\" (UniqueName: \"kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr\") pod \"nova-cell0-db-create-tfqlt\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.845122 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrfsm\" (UniqueName: \"kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.845188 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.845217 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.845324 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhgsl\" (UniqueName: \"kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.849351 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.864875 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.877596 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-053e-account-create-2hclc"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.878697 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.882297 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.897222 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-053e-account-create-2hclc"] Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947341 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947397 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8tzq\" (UniqueName: \"kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947444 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhgsl\" (UniqueName: \"kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947486 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrfsm\" (UniqueName: \"kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947514 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.947538 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.948292 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.949288 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.969724 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhgsl\" (UniqueName: \"kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl\") pod \"nova-cell0-9849-account-create-nkhkx\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:36 crc kubenswrapper[4970]: I1124 14:03:36.984183 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrfsm\" (UniqueName: \"kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm\") pod \"nova-cell1-db-create-b5b2b\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.034120 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.049315 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8tzq\" (UniqueName: \"kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.049588 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.050561 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.057166 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.067356 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8tzq\" (UniqueName: \"kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq\") pod \"nova-cell1-053e-account-create-2hclc\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.204048 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.423399 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:37 crc kubenswrapper[4970]: I1124 14:03:37.424307 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cd4546cfb-h57bs" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.167692 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.308481 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.308815 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.308903 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.308969 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.309047 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.309073 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.309090 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79jbc\" (UniqueName: \"kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc\") pod \"9cfd8b77-9869-4672-865e-e39e8a752567\" (UID: \"9cfd8b77-9869-4672-865e-e39e8a752567\") " Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.311599 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.311819 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.315818 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc" (OuterVolumeSpecName: "kube-api-access-79jbc") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "kube-api-access-79jbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.324591 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts" (OuterVolumeSpecName: "scripts") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.361340 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.376460 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0c4f17a4-9a8d-4824-8370-6aea04d166ae","Type":"ContainerStarted","Data":"cc21900dfc82fadf4261ed049f44e27fede2798ad66a1015b0db16fa44591143"} Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.384792 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cfd8b77-9869-4672-865e-e39e8a752567","Type":"ContainerDied","Data":"52b600c1a4e9932b2e359340548bbf049b4b4cf0c5810af6f752f1f8a1a3b8b2"} Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.384840 4970 scope.go:117] "RemoveContainer" containerID="d4d7a2a412199f99b3bb9871a1e7c38164321d367a99cd85aef568645b701844" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.384868 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.407612 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.66625632 podStartE2EDuration="14.407590279s" podCreationTimestamp="2025-11-24 14:03:26 +0000 UTC" firstStartedPulling="2025-11-24 14:03:27.065085835 +0000 UTC m=+1022.352843128" lastFinishedPulling="2025-11-24 14:03:39.806419794 +0000 UTC m=+1035.094177087" observedRunningTime="2025-11-24 14:03:40.398423337 +0000 UTC m=+1035.686180650" watchObservedRunningTime="2025-11-24 14:03:40.407590279 +0000 UTC m=+1035.695347572" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.411683 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.411713 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79jbc\" (UniqueName: \"kubernetes.io/projected/9cfd8b77-9869-4672-865e-e39e8a752567-kube-api-access-79jbc\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.411727 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.411735 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cfd8b77-9869-4672-865e-e39e8a752567-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.411743 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.421861 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.426909 4970 scope.go:117] "RemoveContainer" containerID="3bfa59ab1ab3da913b632d7477d16cf65420809b8271e67474c9200693241ffc" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.439264 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data" (OuterVolumeSpecName: "config-data") pod "9cfd8b77-9869-4672-865e-e39e8a752567" (UID: "9cfd8b77-9869-4672-865e-e39e8a752567"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.449265 4970 scope.go:117] "RemoveContainer" containerID="b60e14d93f47cdb012466987bf39667da80b69067d008c971b7a3ef4db1597d7" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.468721 4970 scope.go:117] "RemoveContainer" containerID="ab0e7b33253a97c28a38e45214fe61a4451b9340d1723954954ac1022788f0ad" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.513585 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.513611 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cfd8b77-9869-4672-865e-e39e8a752567-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:40 crc kubenswrapper[4970]: W1124 14:03:40.593240 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5f4e454_4f77_46f3_8a28_feb38fc462e6.slice/crio-dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577 WatchSource:0}: Error finding container dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577: Status 404 returned error can't find the container with id dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577 Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.593246 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-053e-account-create-2hclc"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.774233 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:40 crc kubenswrapper[4970]: W1124 14:03:40.813779 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2def0468_8470_4258_b55a_b640a71e248c.slice/crio-8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3 WatchSource:0}: Error finding container 8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3: Status 404 returned error can't find the container with id 8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3 Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.813826 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.832611 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:40 crc kubenswrapper[4970]: E1124 14:03:40.833081 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-central-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833097 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-central-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: E1124 14:03:40.833134 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="sg-core" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833141 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="sg-core" Nov 24 14:03:40 crc kubenswrapper[4970]: E1124 14:03:40.833151 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-notification-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833158 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-notification-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: E1124 14:03:40.833169 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833174 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833613 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833636 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-central-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833648 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="ceilometer-notification-agent" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.833657 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="sg-core" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.835267 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.838563 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.839972 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.863459 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b5b2b"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.880121 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-tfqlt"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.895247 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-84e1-account-create-f27tz"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.905819 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hbxpk"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.921126 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927704 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927796 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927829 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927855 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm5cj\" (UniqueName: \"kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927926 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.927989 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.928016 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:40 crc kubenswrapper[4970]: I1124 14:03:40.937675 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9849-account-create-nkhkx"] Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029269 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029307 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029387 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029429 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029446 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029466 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5cj\" (UniqueName: \"kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.029496 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.030631 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.030931 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.035115 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.035360 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.036543 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.037616 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.053776 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm5cj\" (UniqueName: \"kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj\") pod \"ceilometer-0\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.283981 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.410110 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-84e1-account-create-f27tz" event={"ID":"8bf2cf6d-0109-4c10-bc97-0efba40e772f","Type":"ContainerStarted","Data":"b41eee2584d4780cb068ef7889d33e01cc2429fe66f27fc9ce058e04290353c9"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.410351 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-84e1-account-create-f27tz" event={"ID":"8bf2cf6d-0109-4c10-bc97-0efba40e772f","Type":"ContainerStarted","Data":"023cdfa77f8ceab2a0390ef0a336053ece18247eaf9f507f856ec983f167a8f8"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.430686 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-84e1-account-create-f27tz" podStartSLOduration=5.430672688 podStartE2EDuration="5.430672688s" podCreationTimestamp="2025-11-24 14:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:41.426810966 +0000 UTC m=+1036.714568259" watchObservedRunningTime="2025-11-24 14:03:41.430672688 +0000 UTC m=+1036.718429981" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.455139 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfqlt" event={"ID":"2def0468-8470-4258-b55a-b640a71e248c","Type":"ContainerStarted","Data":"5d0e351c82fb3f11bec7386180873a71cde3882a0bed300d3ef74bdd147e48cb"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.455194 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfqlt" event={"ID":"2def0468-8470-4258-b55a-b640a71e248c","Type":"ContainerStarted","Data":"8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.462663 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9849-account-create-nkhkx" event={"ID":"5419ff1f-e0e4-453a-97e1-a3d5651077c7","Type":"ContainerStarted","Data":"c5d334c1a2f85008f81f76bd9aa00ec9c8379feeadf60e05f9dfb1cebdc0e0fd"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.462722 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9849-account-create-nkhkx" event={"ID":"5419ff1f-e0e4-453a-97e1-a3d5651077c7","Type":"ContainerStarted","Data":"a5ac6325ecc639eb3b63f324b07d3b416853eb624ac41fb260b60f2309b2d5b2"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.467524 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5b2b" event={"ID":"8b055400-56b2-41b3-8ecd-2f4235608a51","Type":"ContainerStarted","Data":"996234b67812d8b08e1437234a253f4f1349e9a9d2dfbf0921407f9c13a70412"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.467568 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5b2b" event={"ID":"8b055400-56b2-41b3-8ecd-2f4235608a51","Type":"ContainerStarted","Data":"2fa1b1e040c7c964e543d8b79765e19a23d735c1f49b2ace0cc16e47cab8f0af"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.495403 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-tfqlt" podStartSLOduration=5.495380943 podStartE2EDuration="5.495380943s" podCreationTimestamp="2025-11-24 14:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:41.486249322 +0000 UTC m=+1036.774006635" watchObservedRunningTime="2025-11-24 14:03:41.495380943 +0000 UTC m=+1036.783138226" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.589477 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-b5b2b" podStartSLOduration=5.589456448 podStartE2EDuration="5.589456448s" podCreationTimestamp="2025-11-24 14:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:41.535444916 +0000 UTC m=+1036.823202209" watchObservedRunningTime="2025-11-24 14:03:41.589456448 +0000 UTC m=+1036.877213771" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.598632 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" path="/var/lib/kubelet/pods/9cfd8b77-9869-4672-865e-e39e8a752567/volumes" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.600345 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053e-account-create-2hclc" event={"ID":"d5f4e454-4f77-46f3-8a28-feb38fc462e6","Type":"ContainerStarted","Data":"43d8a1437b02f626bcd750f61e27e686098703af06c0141791dd0bf25ac70fea"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.600367 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053e-account-create-2hclc" event={"ID":"d5f4e454-4f77-46f3-8a28-feb38fc462e6","Type":"ContainerStarted","Data":"dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.600377 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hbxpk" event={"ID":"37faa061-9cb6-4364-9bf1-7cee945a3859","Type":"ContainerStarted","Data":"ccbf580b445ad9bf3c460a99f20117ec6ca78bb0de719c03ced2441378335df8"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.600393 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hbxpk" event={"ID":"37faa061-9cb6-4364-9bf1-7cee945a3859","Type":"ContainerStarted","Data":"f02222dd0df65237608ead24c130677b335b8a1718ec6ac013e8aad0d92ac910"} Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.709458 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c5f4c5bbd-xvgxp" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.709602 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.725905 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-053e-account-create-2hclc" podStartSLOduration=5.725882614 podStartE2EDuration="5.725882614s" podCreationTimestamp="2025-11-24 14:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:41.585737119 +0000 UTC m=+1036.873494412" watchObservedRunningTime="2025-11-24 14:03:41.725882614 +0000 UTC m=+1037.013639907" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.741131 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-hbxpk" podStartSLOduration=5.741109117 podStartE2EDuration="5.741109117s" podCreationTimestamp="2025-11-24 14:03:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:41.613884875 +0000 UTC m=+1036.901642168" watchObservedRunningTime="2025-11-24 14:03:41.741109117 +0000 UTC m=+1037.028866410" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.742996 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:41 crc kubenswrapper[4970]: I1124 14:03:41.749203 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-76bd896ccf-szjxt" Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.006693 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:42 crc kubenswrapper[4970]: W1124 14:03:42.016670 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c8154a3_fad0_493d_b810_47249ab149f2.slice/crio-0b49dc7333ddb15aa2f39347cde35c35edca4729c5c679ed66462a8d9ebabb5d WatchSource:0}: Error finding container 0b49dc7333ddb15aa2f39347cde35c35edca4729c5c679ed66462a8d9ebabb5d: Status 404 returned error can't find the container with id 0b49dc7333ddb15aa2f39347cde35c35edca4729c5c679ed66462a8d9ebabb5d Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.452760 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.453326 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-log" containerID="cri-o://9686e90d1b07bb6994e2237f076190b5e98f06e7872b555e81d33f9e8d2cba37" gracePeriod=30 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.453401 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-httpd" containerID="cri-o://76848b4bd58a805945b82a0892e7891f7816c7970ca978322f2e7d31e9d94a2b" gracePeriod=30 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.546357 4970 generic.go:334] "Generic (PLEG): container finished" podID="2def0468-8470-4258-b55a-b640a71e248c" containerID="5d0e351c82fb3f11bec7386180873a71cde3882a0bed300d3ef74bdd147e48cb" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.546447 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfqlt" event={"ID":"2def0468-8470-4258-b55a-b640a71e248c","Type":"ContainerDied","Data":"5d0e351c82fb3f11bec7386180873a71cde3882a0bed300d3ef74bdd147e48cb"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.548663 4970 generic.go:334] "Generic (PLEG): container finished" podID="5419ff1f-e0e4-453a-97e1-a3d5651077c7" containerID="c5d334c1a2f85008f81f76bd9aa00ec9c8379feeadf60e05f9dfb1cebdc0e0fd" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.548746 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9849-account-create-nkhkx" event={"ID":"5419ff1f-e0e4-453a-97e1-a3d5651077c7","Type":"ContainerDied","Data":"c5d334c1a2f85008f81f76bd9aa00ec9c8379feeadf60e05f9dfb1cebdc0e0fd"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.553570 4970 generic.go:334] "Generic (PLEG): container finished" podID="8b055400-56b2-41b3-8ecd-2f4235608a51" containerID="996234b67812d8b08e1437234a253f4f1349e9a9d2dfbf0921407f9c13a70412" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.553746 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5b2b" event={"ID":"8b055400-56b2-41b3-8ecd-2f4235608a51","Type":"ContainerDied","Data":"996234b67812d8b08e1437234a253f4f1349e9a9d2dfbf0921407f9c13a70412"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.556919 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerStarted","Data":"0b49dc7333ddb15aa2f39347cde35c35edca4729c5c679ed66462a8d9ebabb5d"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.559401 4970 generic.go:334] "Generic (PLEG): container finished" podID="d5f4e454-4f77-46f3-8a28-feb38fc462e6" containerID="43d8a1437b02f626bcd750f61e27e686098703af06c0141791dd0bf25ac70fea" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.559495 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053e-account-create-2hclc" event={"ID":"d5f4e454-4f77-46f3-8a28-feb38fc462e6","Type":"ContainerDied","Data":"43d8a1437b02f626bcd750f61e27e686098703af06c0141791dd0bf25ac70fea"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.582144 4970 generic.go:334] "Generic (PLEG): container finished" podID="37faa061-9cb6-4364-9bf1-7cee945a3859" containerID="ccbf580b445ad9bf3c460a99f20117ec6ca78bb0de719c03ced2441378335df8" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.582287 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hbxpk" event={"ID":"37faa061-9cb6-4364-9bf1-7cee945a3859","Type":"ContainerDied","Data":"ccbf580b445ad9bf3c460a99f20117ec6ca78bb0de719c03ced2441378335df8"} Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.584374 4970 generic.go:334] "Generic (PLEG): container finished" podID="8bf2cf6d-0109-4c10-bc97-0efba40e772f" containerID="b41eee2584d4780cb068ef7889d33e01cc2429fe66f27fc9ce058e04290353c9" exitCode=0 Nov 24 14:03:42 crc kubenswrapper[4970]: I1124 14:03:42.584599 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-84e1-account-create-f27tz" event={"ID":"8bf2cf6d-0109-4c10-bc97-0efba40e772f","Type":"ContainerDied","Data":"b41eee2584d4780cb068ef7889d33e01cc2429fe66f27fc9ce058e04290353c9"} Nov 24 14:03:43 crc kubenswrapper[4970]: I1124 14:03:43.599733 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerStarted","Data":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} Nov 24 14:03:43 crc kubenswrapper[4970]: I1124 14:03:43.600251 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerStarted","Data":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} Nov 24 14:03:43 crc kubenswrapper[4970]: I1124 14:03:43.602926 4970 generic.go:334] "Generic (PLEG): container finished" podID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerID="9686e90d1b07bb6994e2237f076190b5e98f06e7872b555e81d33f9e8d2cba37" exitCode=143 Nov 24 14:03:43 crc kubenswrapper[4970]: I1124 14:03:43.603020 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerDied","Data":"9686e90d1b07bb6994e2237f076190b5e98f06e7872b555e81d33f9e8d2cba37"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.041225 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.076319 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.076673 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-log" containerID="cri-o://8569b00245973b27500b9015ba1ec080a6f3671a33e1a590c81ad8c215dfdbf1" gracePeriod=30 Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.076952 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-httpd" containerID="cri-o://162e41175b6f5d6fb2f700370057743c2e9dbb35c5f9e3ab38d4c2b4c1e2e410" gracePeriod=30 Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.153206 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.158390 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.224311 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts\") pod \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.224381 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8tzq\" (UniqueName: \"kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq\") pod \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\" (UID: \"d5f4e454-4f77-46f3-8a28-feb38fc462e6\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.224903 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5f4e454-4f77-46f3-8a28-feb38fc462e6" (UID: "d5f4e454-4f77-46f3-8a28-feb38fc462e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.230877 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq" (OuterVolumeSpecName: "kube-api-access-r8tzq") pod "d5f4e454-4f77-46f3-8a28-feb38fc462e6" (UID: "d5f4e454-4f77-46f3-8a28-feb38fc462e6"). InnerVolumeSpecName "kube-api-access-r8tzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.325999 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57g6r\" (UniqueName: \"kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r\") pod \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.326049 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts\") pod \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\" (UID: \"8bf2cf6d-0109-4c10-bc97-0efba40e772f\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.326066 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhgsl\" (UniqueName: \"kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl\") pod \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.326176 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts\") pod \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\" (UID: \"5419ff1f-e0e4-453a-97e1-a3d5651077c7\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.326545 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8tzq\" (UniqueName: \"kubernetes.io/projected/d5f4e454-4f77-46f3-8a28-feb38fc462e6-kube-api-access-r8tzq\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.326558 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5f4e454-4f77-46f3-8a28-feb38fc462e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.328100 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8bf2cf6d-0109-4c10-bc97-0efba40e772f" (UID: "8bf2cf6d-0109-4c10-bc97-0efba40e772f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.329838 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5419ff1f-e0e4-453a-97e1-a3d5651077c7" (UID: "5419ff1f-e0e4-453a-97e1-a3d5651077c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.333769 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r" (OuterVolumeSpecName: "kube-api-access-57g6r") pod "8bf2cf6d-0109-4c10-bc97-0efba40e772f" (UID: "8bf2cf6d-0109-4c10-bc97-0efba40e772f"). InnerVolumeSpecName "kube-api-access-57g6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.352834 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl" (OuterVolumeSpecName: "kube-api-access-lhgsl") pod "5419ff1f-e0e4-453a-97e1-a3d5651077c7" (UID: "5419ff1f-e0e4-453a-97e1-a3d5651077c7"). InnerVolumeSpecName "kube-api-access-lhgsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.429671 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5419ff1f-e0e4-453a-97e1-a3d5651077c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.429708 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57g6r\" (UniqueName: \"kubernetes.io/projected/8bf2cf6d-0109-4c10-bc97-0efba40e772f-kube-api-access-57g6r\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.429720 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8bf2cf6d-0109-4c10-bc97-0efba40e772f-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.429730 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhgsl\" (UniqueName: \"kubernetes.io/projected/5419ff1f-e0e4-453a-97e1-a3d5651077c7-kube-api-access-lhgsl\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.482268 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.488967 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.495510 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.614652 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-053e-account-create-2hclc" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.615882 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-053e-account-create-2hclc" event={"ID":"d5f4e454-4f77-46f3-8a28-feb38fc462e6","Type":"ContainerDied","Data":"dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.615922 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dadd542c6c830fcc91e5a95e32ae278a54dc248b8fa646e80fbf1674d66e7577" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.617622 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hbxpk" event={"ID":"37faa061-9cb6-4364-9bf1-7cee945a3859","Type":"ContainerDied","Data":"f02222dd0df65237608ead24c130677b335b8a1718ec6ac013e8aad0d92ac910"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.617650 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f02222dd0df65237608ead24c130677b335b8a1718ec6ac013e8aad0d92ac910" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.617701 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hbxpk" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.620178 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b5b2b" event={"ID":"8b055400-56b2-41b3-8ecd-2f4235608a51","Type":"ContainerDied","Data":"2fa1b1e040c7c964e543d8b79765e19a23d735c1f49b2ace0cc16e47cab8f0af"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.620210 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fa1b1e040c7c964e543d8b79765e19a23d735c1f49b2ace0cc16e47cab8f0af" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.620262 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b5b2b" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.622133 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-84e1-account-create-f27tz" event={"ID":"8bf2cf6d-0109-4c10-bc97-0efba40e772f","Type":"ContainerDied","Data":"023cdfa77f8ceab2a0390ef0a336053ece18247eaf9f507f856ec983f167a8f8"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.622153 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="023cdfa77f8ceab2a0390ef0a336053ece18247eaf9f507f856ec983f167a8f8" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.622188 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-84e1-account-create-f27tz" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635336 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts\") pod \"2def0468-8470-4258-b55a-b640a71e248c\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635456 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrfsm\" (UniqueName: \"kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm\") pod \"8b055400-56b2-41b3-8ecd-2f4235608a51\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635682 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts\") pod \"8b055400-56b2-41b3-8ecd-2f4235608a51\" (UID: \"8b055400-56b2-41b3-8ecd-2f4235608a51\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635724 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vhlr\" (UniqueName: \"kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr\") pod \"2def0468-8470-4258-b55a-b640a71e248c\" (UID: \"2def0468-8470-4258-b55a-b640a71e248c\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635771 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q87xl\" (UniqueName: \"kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl\") pod \"37faa061-9cb6-4364-9bf1-7cee945a3859\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.635808 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts\") pod \"37faa061-9cb6-4364-9bf1-7cee945a3859\" (UID: \"37faa061-9cb6-4364-9bf1-7cee945a3859\") " Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.636042 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b055400-56b2-41b3-8ecd-2f4235608a51" (UID: "8b055400-56b2-41b3-8ecd-2f4235608a51"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.636345 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b055400-56b2-41b3-8ecd-2f4235608a51-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.636618 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerStarted","Data":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.636635 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37faa061-9cb6-4364-9bf1-7cee945a3859" (UID: "37faa061-9cb6-4364-9bf1-7cee945a3859"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.636747 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2def0468-8470-4258-b55a-b640a71e248c" (UID: "2def0468-8470-4258-b55a-b640a71e248c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.640218 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm" (OuterVolumeSpecName: "kube-api-access-jrfsm") pod "8b055400-56b2-41b3-8ecd-2f4235608a51" (UID: "8b055400-56b2-41b3-8ecd-2f4235608a51"). InnerVolumeSpecName "kube-api-access-jrfsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.640289 4970 generic.go:334] "Generic (PLEG): container finished" podID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerID="8569b00245973b27500b9015ba1ec080a6f3671a33e1a590c81ad8c215dfdbf1" exitCode=143 Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.640367 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerDied","Data":"8569b00245973b27500b9015ba1ec080a6f3671a33e1a590c81ad8c215dfdbf1"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.641952 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr" (OuterVolumeSpecName: "kube-api-access-9vhlr") pod "2def0468-8470-4258-b55a-b640a71e248c" (UID: "2def0468-8470-4258-b55a-b640a71e248c"). InnerVolumeSpecName "kube-api-access-9vhlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.643247 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-tfqlt" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.643428 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-tfqlt" event={"ID":"2def0468-8470-4258-b55a-b640a71e248c","Type":"ContainerDied","Data":"8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.643496 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a40de0ce40f721ecb3a02ef7c3e53f74ccae7fcd32a036b1b1b0051827c66a3" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.645256 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9849-account-create-nkhkx" event={"ID":"5419ff1f-e0e4-453a-97e1-a3d5651077c7","Type":"ContainerDied","Data":"a5ac6325ecc639eb3b63f324b07d3b416853eb624ac41fb260b60f2309b2d5b2"} Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.645277 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9849-account-create-nkhkx" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.645279 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5ac6325ecc639eb3b63f324b07d3b416853eb624ac41fb260b60f2309b2d5b2" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.654733 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl" (OuterVolumeSpecName: "kube-api-access-q87xl") pod "37faa061-9cb6-4364-9bf1-7cee945a3859" (UID: "37faa061-9cb6-4364-9bf1-7cee945a3859"). InnerVolumeSpecName "kube-api-access-q87xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.722081 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.737893 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vhlr\" (UniqueName: \"kubernetes.io/projected/2def0468-8470-4258-b55a-b640a71e248c-kube-api-access-9vhlr\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.737934 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q87xl\" (UniqueName: \"kubernetes.io/projected/37faa061-9cb6-4364-9bf1-7cee945a3859-kube-api-access-q87xl\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.737948 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37faa061-9cb6-4364-9bf1-7cee945a3859-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.737959 4970 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2def0468-8470-4258-b55a-b640a71e248c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:44 crc kubenswrapper[4970]: I1124 14:03:44.737973 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrfsm\" (UniqueName: \"kubernetes.io/projected/8b055400-56b2-41b3-8ecd-2f4235608a51-kube-api-access-jrfsm\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:46 crc kubenswrapper[4970]: I1124 14:03:46.666746 4970 generic.go:334] "Generic (PLEG): container finished" podID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerID="76848b4bd58a805945b82a0892e7891f7816c7970ca978322f2e7d31e9d94a2b" exitCode=0 Nov 24 14:03:46 crc kubenswrapper[4970]: I1124 14:03:46.666834 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerDied","Data":"76848b4bd58a805945b82a0892e7891f7816c7970ca978322f2e7d31e9d94a2b"} Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.062930 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7mjd6"] Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063623 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37faa061-9cb6-4364-9bf1-7cee945a3859" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063644 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="37faa061-9cb6-4364-9bf1-7cee945a3859" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063670 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2def0468-8470-4258-b55a-b640a71e248c" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063679 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2def0468-8470-4258-b55a-b640a71e248c" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063693 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5419ff1f-e0e4-453a-97e1-a3d5651077c7" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063704 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5419ff1f-e0e4-453a-97e1-a3d5651077c7" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063716 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b055400-56b2-41b3-8ecd-2f4235608a51" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063723 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b055400-56b2-41b3-8ecd-2f4235608a51" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063739 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf2cf6d-0109-4c10-bc97-0efba40e772f" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063747 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf2cf6d-0109-4c10-bc97-0efba40e772f" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.063770 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f4e454-4f77-46f3-8a28-feb38fc462e6" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063778 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f4e454-4f77-46f3-8a28-feb38fc462e6" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.063993 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f4e454-4f77-46f3-8a28-feb38fc462e6" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064009 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b055400-56b2-41b3-8ecd-2f4235608a51" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064031 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5419ff1f-e0e4-453a-97e1-a3d5651077c7" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064043 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf2cf6d-0109-4c10-bc97-0efba40e772f" containerName="mariadb-account-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064057 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="2def0468-8470-4258-b55a-b640a71e248c" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064066 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="37faa061-9cb6-4364-9bf1-7cee945a3859" containerName="mariadb-database-create" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.064885 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.068274 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.069346 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jt59z" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.069466 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.073715 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7mjd6"] Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.177830 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.177913 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.177967 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.177991 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jcjs\" (UniqueName: \"kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.221535 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.282164 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.282219 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.282267 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.282292 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jcjs\" (UniqueName: \"kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.289892 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.291270 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.292426 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.298592 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jcjs\" (UniqueName: \"kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs\") pod \"nova-cell0-conductor-db-sync-7mjd6\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.383508 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.383908 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.383944 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zgq5\" (UniqueName: \"kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.383996 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.384033 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.384057 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.384081 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.384111 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7455c95f-d324-47bd-86f2-2ed4dee790c0\" (UID: \"7455c95f-d324-47bd-86f2-2ed4dee790c0\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.385961 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.386739 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs" (OuterVolumeSpecName: "logs") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.389867 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.392767 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.397734 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts" (OuterVolumeSpecName: "scripts") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.407560 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5" (OuterVolumeSpecName: "kube-api-access-5zgq5") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "kube-api-access-5zgq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.431797 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.460588 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487792 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487827 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487839 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zgq5\" (UniqueName: \"kubernetes.io/projected/7455c95f-d324-47bd-86f2-2ed4dee790c0-kube-api-access-5zgq5\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487850 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487863 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7455c95f-d324-47bd-86f2-2ed4dee790c0-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487873 4970 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.487898 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.506721 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data" (OuterVolumeSpecName: "config-data") pod "7455c95f-d324-47bd-86f2-2ed4dee790c0" (UID: "7455c95f-d324-47bd-86f2-2ed4dee790c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.530486 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.589561 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7455c95f-d324-47bd-86f2-2ed4dee790c0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.589620 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.687854 4970 generic.go:334] "Generic (PLEG): container finished" podID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerID="ca522efedbf9731e67c7b3e65f8e485ba5ff868b422ba102b15ddfee9e01717d" exitCode=137 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.687964 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerDied","Data":"ca522efedbf9731e67c7b3e65f8e485ba5ff868b422ba102b15ddfee9e01717d"} Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.692283 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7455c95f-d324-47bd-86f2-2ed4dee790c0","Type":"ContainerDied","Data":"2b231424689eda4ae371e87749a89d5bdd1337f8d159e5622d8787552270d28d"} Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.692322 4970 scope.go:117] "RemoveContainer" containerID="76848b4bd58a805945b82a0892e7891f7816c7970ca978322f2e7d31e9d94a2b" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.692439 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.696857 4970 generic.go:334] "Generic (PLEG): container finished" podID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerID="162e41175b6f5d6fb2f700370057743c2e9dbb35c5f9e3ab38d4c2b4c1e2e410" exitCode=0 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.696956 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerDied","Data":"162e41175b6f5d6fb2f700370057743c2e9dbb35c5f9e3ab38d4c2b4c1e2e410"} Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.699672 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerStarted","Data":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.699886 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-central-agent" containerID="cri-o://2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" gracePeriod=30 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.700150 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="sg-core" containerID="cri-o://168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" gracePeriod=30 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.700170 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.700360 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-notification-agent" containerID="cri-o://fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" gracePeriod=30 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.700491 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="proxy-httpd" containerID="cri-o://87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" gracePeriod=30 Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.732254 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.078387805 podStartE2EDuration="7.732211805s" podCreationTimestamp="2025-11-24 14:03:40 +0000 UTC" firstStartedPulling="2025-11-24 14:03:42.020923385 +0000 UTC m=+1037.308680678" lastFinishedPulling="2025-11-24 14:03:46.674747385 +0000 UTC m=+1041.962504678" observedRunningTime="2025-11-24 14:03:47.726179635 +0000 UTC m=+1043.013936928" watchObservedRunningTime="2025-11-24 14:03:47.732211805 +0000 UTC m=+1043.019969098" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.821210 4970 scope.go:117] "RemoveContainer" containerID="9686e90d1b07bb6994e2237f076190b5e98f06e7872b555e81d33f9e8d2cba37" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.835308 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.844831 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.873498 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.905312 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.905840 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon-log" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.905859 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon-log" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.905877 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-log" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.905885 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-log" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.905896 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.905904 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" Nov 24 14:03:47 crc kubenswrapper[4970]: E1124 14:03:47.905926 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-httpd" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.905935 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-httpd" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.906152 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-log" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.906176 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" containerName="glance-httpd" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.906194 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon-log" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.906211 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" containerName="horizon" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.907384 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.921010 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.921372 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.961673 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.974108 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7mjd6"] Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996128 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996348 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996420 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996446 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996521 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996779 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx7xb\" (UniqueName: \"kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:47 crc kubenswrapper[4970]: I1124 14:03:47.996976 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs\") pod \"821ffced-26a3-4ad0-a040-7145f772ce5c\" (UID: \"821ffced-26a3-4ad0-a040-7145f772ce5c\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998051 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnkmw\" (UniqueName: \"kubernetes.io/projected/0713092e-39d2-40e1-b708-a2e2addd8249-kube-api-access-hnkmw\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998115 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-logs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998159 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-config-data\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998216 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998558 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998694 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998728 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-scripts\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:47.998958 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.003258 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.003780 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs" (OuterVolumeSpecName: "logs") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.016849 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb" (OuterVolumeSpecName: "kube-api-access-xx7xb") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "kube-api-access-xx7xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.031469 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data" (OuterVolumeSpecName: "config-data") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.045783 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.047726 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts" (OuterVolumeSpecName: "scripts") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.051655 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100277 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100311 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100338 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-scripts\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100371 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100409 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnkmw\" (UniqueName: \"kubernetes.io/projected/0713092e-39d2-40e1-b708-a2e2addd8249-kube-api-access-hnkmw\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100425 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-logs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100453 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-config-data\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100476 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100523 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx7xb\" (UniqueName: \"kubernetes.io/projected/821ffced-26a3-4ad0-a040-7145f772ce5c-kube-api-access-xx7xb\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100535 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/821ffced-26a3-4ad0-a040-7145f772ce5c-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100543 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100551 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/821ffced-26a3-4ad0-a040-7145f772ce5c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100559 4970 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100567 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.100743 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.104251 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.105077 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0713092e-39d2-40e1-b708-a2e2addd8249-logs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.112143 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-scripts\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.112386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.115950 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.118347 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0713092e-39d2-40e1-b708-a2e2addd8249-config-data\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.130340 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnkmw\" (UniqueName: \"kubernetes.io/projected/0713092e-39d2-40e1-b708-a2e2addd8249-kube-api-access-hnkmw\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.132440 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "821ffced-26a3-4ad0-a040-7145f772ce5c" (UID: "821ffced-26a3-4ad0-a040-7145f772ce5c"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.142735 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0713092e-39d2-40e1-b708-a2e2addd8249\") " pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202312 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202385 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h54mr\" (UniqueName: \"kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202423 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202452 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202476 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202545 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202639 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.202765 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs\") pod \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\" (UID: \"d6c6d650-8efb-4ac5-9ada-64713f74e36a\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.203178 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.203247 4970 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/821ffced-26a3-4ad0-a040-7145f772ce5c-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.204765 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs" (OuterVolumeSpecName: "logs") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.206321 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr" (OuterVolumeSpecName: "kube-api-access-h54mr") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "kube-api-access-h54mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.206521 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts" (OuterVolumeSpecName: "scripts") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.209741 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.230413 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.240671 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.257703 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data" (OuterVolumeSpecName: "config-data") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.271131 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d6c6d650-8efb-4ac5-9ada-64713f74e36a" (UID: "d6c6d650-8efb-4ac5-9ada-64713f74e36a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305271 4970 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305311 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h54mr\" (UniqueName: \"kubernetes.io/projected/d6c6d650-8efb-4ac5-9ada-64713f74e36a-kube-api-access-h54mr\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305326 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305335 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305346 4970 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305355 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c6d650-8efb-4ac5-9ada-64713f74e36a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305389 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.305400 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6c6d650-8efb-4ac5-9ada-64713f74e36a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.332284 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.407853 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.539570 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.610313 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm5cj\" (UniqueName: \"kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.610757 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.610855 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.610943 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.611006 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.611063 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.611192 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts\") pod \"0c8154a3-fad0-493d-b810-47249ab149f2\" (UID: \"0c8154a3-fad0-493d-b810-47249ab149f2\") " Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.611567 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.611706 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.612339 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.617425 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj" (OuterVolumeSpecName: "kube-api-access-cm5cj") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "kube-api-access-cm5cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.617616 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts" (OuterVolumeSpecName: "scripts") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.642389 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.682262 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.702553 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data" (OuterVolumeSpecName: "config-data") pod "0c8154a3-fad0-493d-b810-47249ab149f2" (UID: "0c8154a3-fad0-493d-b810-47249ab149f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713079 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713124 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713137 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713148 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c8154a3-fad0-493d-b810-47249ab149f2-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713161 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8154a3-fad0-493d-b810-47249ab149f2-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.713172 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm5cj\" (UniqueName: \"kubernetes.io/projected/0c8154a3-fad0-493d-b810-47249ab149f2-kube-api-access-cm5cj\") on node \"crc\" DevicePath \"\"" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.714166 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c5f4c5bbd-xvgxp" event={"ID":"821ffced-26a3-4ad0-a040-7145f772ce5c","Type":"ContainerDied","Data":"056363c1e0d4e437d7a1045c2498d225b2ec2f0473410e76f98de4d4e9493418"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.714221 4970 scope.go:117] "RemoveContainer" containerID="6f5a1170f608d84dfefcb3970f5b308afb8b70329afa1e99dcc81638e534855d" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.714372 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c5f4c5bbd-xvgxp" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.722375 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" event={"ID":"a45bf61d-f620-4455-b4b0-6d0973225fdb","Type":"ContainerStarted","Data":"fa9917de1e922225da44da3e12c956f1e3737be1522f80aabeb03841ed149866"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.724924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6c6d650-8efb-4ac5-9ada-64713f74e36a","Type":"ContainerDied","Data":"18743ef26e78757d652cdf4969aa9a5924a012b8f4ce912e98704ff409f00388"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.725015 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729027 4970 generic.go:334] "Generic (PLEG): container finished" podID="0c8154a3-fad0-493d-b810-47249ab149f2" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" exitCode=0 Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729051 4970 generic.go:334] "Generic (PLEG): container finished" podID="0c8154a3-fad0-493d-b810-47249ab149f2" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" exitCode=2 Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729061 4970 generic.go:334] "Generic (PLEG): container finished" podID="0c8154a3-fad0-493d-b810-47249ab149f2" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" exitCode=0 Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729069 4970 generic.go:334] "Generic (PLEG): container finished" podID="0c8154a3-fad0-493d-b810-47249ab149f2" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" exitCode=0 Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729086 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerDied","Data":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729090 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729108 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerDied","Data":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729121 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerDied","Data":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729131 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerDied","Data":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.729142 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c8154a3-fad0-493d-b810-47249ab149f2","Type":"ContainerDied","Data":"0b49dc7333ddb15aa2f39347cde35c35edca4729c5c679ed66462a8d9ebabb5d"} Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.831776 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: W1124 14:03:48.833500 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0713092e_39d2_40e1_b708_a2e2addd8249.slice/crio-476a6be6404cfb7c055a836b8af85dbb706be05dfeb1465458ed652977423f2c WatchSource:0}: Error finding container 476a6be6404cfb7c055a836b8af85dbb706be05dfeb1465458ed652977423f2c: Status 404 returned error can't find the container with id 476a6be6404cfb7c055a836b8af85dbb706be05dfeb1465458ed652977423f2c Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.853040 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.870600 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.886648 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.900889 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.909744 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910244 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-central-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910260 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-central-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910276 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="proxy-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910284 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="proxy-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910297 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-notification-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910305 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-notification-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910333 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="sg-core" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910341 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="sg-core" Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910351 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910358 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: E1124 14:03:48.910372 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-log" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910381 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-log" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910609 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-central-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910622 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-log" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910640 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" containerName="glance-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910657 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="proxy-httpd" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910672 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="ceilometer-notification-agent" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.910685 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" containerName="sg-core" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.911891 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.916177 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.927520 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.927729 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.932799 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.953021 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c5f4c5bbd-xvgxp"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.966870 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.969882 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.969978 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.973941 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:03:48 crc kubenswrapper[4970]: I1124 14:03:48.974131 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024521 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024561 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024625 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024824 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc9gm\" (UniqueName: \"kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024850 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024866 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plv4v\" (UniqueName: \"kubernetes.io/projected/639b382e-fe48-4188-ba6b-1b2c5688898c-kube-api-access-plv4v\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024882 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024896 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024912 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024927 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.024980 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.025003 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-logs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.025017 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.025041 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.025054 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.113759 4970 scope.go:117] "RemoveContainer" containerID="ca522efedbf9731e67c7b3e65f8e485ba5ff868b422ba102b15ddfee9e01717d" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.126306 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc9gm\" (UniqueName: \"kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.126358 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.126387 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plv4v\" (UniqueName: \"kubernetes.io/projected/639b382e-fe48-4188-ba6b-1b2c5688898c-kube-api-access-plv4v\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.129912 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130062 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130100 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130160 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130324 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130378 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-logs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130404 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130450 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.130473 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.131080 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.131147 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.131412 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.132305 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.133035 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.136787 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.137043 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/639b382e-fe48-4188-ba6b-1b2c5688898c-logs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.137239 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.142688 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.148423 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.156010 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.158218 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.164131 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.164231 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.164318 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.164669 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plv4v\" (UniqueName: \"kubernetes.io/projected/639b382e-fe48-4188-ba6b-1b2c5688898c-kube-api-access-plv4v\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.175364 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/639b382e-fe48-4188-ba6b-1b2c5688898c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.181514 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc9gm\" (UniqueName: \"kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm\") pod \"ceilometer-0\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.213045 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"639b382e-fe48-4188-ba6b-1b2c5688898c\") " pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.255637 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.311526 4970 scope.go:117] "RemoveContainer" containerID="162e41175b6f5d6fb2f700370057743c2e9dbb35c5f9e3ab38d4c2b4c1e2e410" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.367142 4970 scope.go:117] "RemoveContainer" containerID="8569b00245973b27500b9015ba1ec080a6f3671a33e1a590c81ad8c215dfdbf1" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.373730 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.433421 4970 scope.go:117] "RemoveContainer" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.509276 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c8154a3-fad0-493d-b810-47249ab149f2" path="/var/lib/kubelet/pods/0c8154a3-fad0-493d-b810-47249ab149f2/volumes" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.510253 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7455c95f-d324-47bd-86f2-2ed4dee790c0" path="/var/lib/kubelet/pods/7455c95f-d324-47bd-86f2-2ed4dee790c0/volumes" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.511990 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="821ffced-26a3-4ad0-a040-7145f772ce5c" path="/var/lib/kubelet/pods/821ffced-26a3-4ad0-a040-7145f772ce5c/volumes" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.512829 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6c6d650-8efb-4ac5-9ada-64713f74e36a" path="/var/lib/kubelet/pods/d6c6d650-8efb-4ac5-9ada-64713f74e36a/volumes" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.520805 4970 scope.go:117] "RemoveContainer" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.558854 4970 scope.go:117] "RemoveContainer" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.584894 4970 scope.go:117] "RemoveContainer" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.616697 4970 scope.go:117] "RemoveContainer" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: E1124 14:03:49.617452 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": container with ID starting with 87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631 not found: ID does not exist" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.617497 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} err="failed to get container status \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": rpc error: code = NotFound desc = could not find container \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": container with ID starting with 87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.617528 4970 scope.go:117] "RemoveContainer" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: E1124 14:03:49.618564 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": container with ID starting with 168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55 not found: ID does not exist" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.618616 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} err="failed to get container status \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": rpc error: code = NotFound desc = could not find container \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": container with ID starting with 168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.618647 4970 scope.go:117] "RemoveContainer" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: E1124 14:03:49.620011 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": container with ID starting with fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509 not found: ID does not exist" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.620058 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} err="failed to get container status \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": rpc error: code = NotFound desc = could not find container \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": container with ID starting with fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.620090 4970 scope.go:117] "RemoveContainer" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: E1124 14:03:49.620502 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": container with ID starting with 2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef not found: ID does not exist" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.620547 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} err="failed to get container status \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": rpc error: code = NotFound desc = could not find container \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": container with ID starting with 2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.620601 4970 scope.go:117] "RemoveContainer" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621008 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} err="failed to get container status \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": rpc error: code = NotFound desc = could not find container \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": container with ID starting with 87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621037 4970 scope.go:117] "RemoveContainer" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621459 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} err="failed to get container status \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": rpc error: code = NotFound desc = could not find container \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": container with ID starting with 168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621474 4970 scope.go:117] "RemoveContainer" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621649 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} err="failed to get container status \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": rpc error: code = NotFound desc = could not find container \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": container with ID starting with fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621662 4970 scope.go:117] "RemoveContainer" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621810 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} err="failed to get container status \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": rpc error: code = NotFound desc = could not find container \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": container with ID starting with 2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.621827 4970 scope.go:117] "RemoveContainer" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.624170 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} err="failed to get container status \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": rpc error: code = NotFound desc = could not find container \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": container with ID starting with 87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.624192 4970 scope.go:117] "RemoveContainer" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.624779 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} err="failed to get container status \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": rpc error: code = NotFound desc = could not find container \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": container with ID starting with 168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.624798 4970 scope.go:117] "RemoveContainer" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.625562 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} err="failed to get container status \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": rpc error: code = NotFound desc = could not find container \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": container with ID starting with fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.625595 4970 scope.go:117] "RemoveContainer" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.625810 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} err="failed to get container status \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": rpc error: code = NotFound desc = could not find container \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": container with ID starting with 2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.625839 4970 scope.go:117] "RemoveContainer" containerID="87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626009 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631"} err="failed to get container status \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": rpc error: code = NotFound desc = could not find container \"87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631\": container with ID starting with 87c525214856a8da0bce59320c435ce43cd4bc2ad6b3d424c5974239d5ea4631 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626028 4970 scope.go:117] "RemoveContainer" containerID="168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626347 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55"} err="failed to get container status \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": rpc error: code = NotFound desc = could not find container \"168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55\": container with ID starting with 168a9f36e42278b32b57c8851c2f91a46ebdcd202275421114bf9acfa354ec55 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626367 4970 scope.go:117] "RemoveContainer" containerID="fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626763 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509"} err="failed to get container status \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": rpc error: code = NotFound desc = could not find container \"fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509\": container with ID starting with fffbc9a6c05d1702cb1afb87587a50b6c657ac4f29f54c4eead86fc50e94c509 not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626781 4970 scope.go:117] "RemoveContainer" containerID="2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.626926 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef"} err="failed to get container status \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": rpc error: code = NotFound desc = could not find container \"2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef\": container with ID starting with 2e8c07e7e3fd0687b09025b5623bdc6ffcb7fa7021728baec473f4e49e3e98ef not found: ID does not exist" Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.753509 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0713092e-39d2-40e1-b708-a2e2addd8249","Type":"ContainerStarted","Data":"476a6be6404cfb7c055a836b8af85dbb706be05dfeb1465458ed652977423f2c"} Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.881931 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 24 14:03:49 crc kubenswrapper[4970]: I1124 14:03:49.992615 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.773664 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"639b382e-fe48-4188-ba6b-1b2c5688898c","Type":"ContainerStarted","Data":"520b0b6905f92234573421880cb2b51d4c09bc440046dd2e51c9207dfba1f32b"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.774025 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"639b382e-fe48-4188-ba6b-1b2c5688898c","Type":"ContainerStarted","Data":"19ca7f98bf6a7e0cc3133405c83863d0030c5a223edd5230f4bf9eb71780e949"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.775509 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerStarted","Data":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.775538 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerStarted","Data":"8d3910788c08f4072d5bb7a13a9d42080a56b869003a4f5e0afdfe46ebb77114"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.778155 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0713092e-39d2-40e1-b708-a2e2addd8249","Type":"ContainerStarted","Data":"cfa1febc5abde412100e44a5d8a14757765aee711e660da2842aaffa970aff9e"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.778183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0713092e-39d2-40e1-b708-a2e2addd8249","Type":"ContainerStarted","Data":"75dd2e4aa0e950610b9b29230ddebb0656b5fb93440124a1f9232a89d6ec4308"} Nov 24 14:03:50 crc kubenswrapper[4970]: I1124 14:03:50.797490 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.797472526 podStartE2EDuration="3.797472526s" podCreationTimestamp="2025-11-24 14:03:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:50.795762181 +0000 UTC m=+1046.083519474" watchObservedRunningTime="2025-11-24 14:03:50.797472526 +0000 UTC m=+1046.085229819" Nov 24 14:03:51 crc kubenswrapper[4970]: I1124 14:03:51.138976 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:03:51 crc kubenswrapper[4970]: I1124 14:03:51.792250 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerStarted","Data":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} Nov 24 14:03:51 crc kubenswrapper[4970]: I1124 14:03:51.795679 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"639b382e-fe48-4188-ba6b-1b2c5688898c","Type":"ContainerStarted","Data":"84de491a61542840a890edd80fa3b317cf0e8d34d2bf20e1d45ac9cea2c3ce43"} Nov 24 14:03:51 crc kubenswrapper[4970]: I1124 14:03:51.813019 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.813002924 podStartE2EDuration="3.813002924s" podCreationTimestamp="2025-11-24 14:03:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:03:51.812612894 +0000 UTC m=+1047.100370187" watchObservedRunningTime="2025-11-24 14:03:51.813002924 +0000 UTC m=+1047.100760217" Nov 24 14:03:52 crc kubenswrapper[4970]: I1124 14:03:52.811706 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerStarted","Data":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.240877 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.241375 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.294130 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.294370 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.867684 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" event={"ID":"a45bf61d-f620-4455-b4b0-6d0973225fdb","Type":"ContainerStarted","Data":"7c43ed5de706ff312a95ad504212c825b98a056620129c5b1ae8b01e5d1d16b9"} Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871472 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerStarted","Data":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871724 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871767 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871726 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-central-agent" containerID="cri-o://4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" gracePeriod=30 Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871786 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="sg-core" containerID="cri-o://dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" gracePeriod=30 Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871813 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="proxy-httpd" containerID="cri-o://501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" gracePeriod=30 Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.871803 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-notification-agent" containerID="cri-o://19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" gracePeriod=30 Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.896355 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" podStartSLOduration=2.018388927 podStartE2EDuration="11.896339254s" podCreationTimestamp="2025-11-24 14:03:47 +0000 UTC" firstStartedPulling="2025-11-24 14:03:47.953270684 +0000 UTC m=+1043.241027977" lastFinishedPulling="2025-11-24 14:03:57.831221011 +0000 UTC m=+1053.118978304" observedRunningTime="2025-11-24 14:03:58.888711302 +0000 UTC m=+1054.176468625" watchObservedRunningTime="2025-11-24 14:03:58.896339254 +0000 UTC m=+1054.184096547" Nov 24 14:03:58 crc kubenswrapper[4970]: I1124 14:03:58.933454 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.12583892 podStartE2EDuration="10.933435598s" podCreationTimestamp="2025-11-24 14:03:48 +0000 UTC" firstStartedPulling="2025-11-24 14:03:50.008149853 +0000 UTC m=+1045.295907156" lastFinishedPulling="2025-11-24 14:03:57.815746541 +0000 UTC m=+1053.103503834" observedRunningTime="2025-11-24 14:03:58.923658029 +0000 UTC m=+1054.211415332" watchObservedRunningTime="2025-11-24 14:03:58.933435598 +0000 UTC m=+1054.221192891" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.256715 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.256772 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.289125 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.312222 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.836458 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.884905 4970 generic.go:334] "Generic (PLEG): container finished" podID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" exitCode=0 Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.884944 4970 generic.go:334] "Generic (PLEG): container finished" podID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" exitCode=2 Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.884956 4970 generic.go:334] "Generic (PLEG): container finished" podID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" exitCode=0 Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.884965 4970 generic.go:334] "Generic (PLEG): container finished" podID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" exitCode=0 Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.884959 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerDied","Data":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885016 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerDied","Data":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885027 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerDied","Data":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885037 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerDied","Data":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885046 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ea5a414-ff51-4622-8c07-4e1bf562d4c8","Type":"ContainerDied","Data":"8d3910788c08f4072d5bb7a13a9d42080a56b869003a4f5e0afdfe46ebb77114"} Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885061 4970 scope.go:117] "RemoveContainer" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885004 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885349 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.885390 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.916789 4970 scope.go:117] "RemoveContainer" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.949738 4970 scope.go:117] "RemoveContainer" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.974985 4970 scope.go:117] "RemoveContainer" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981600 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981670 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981775 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981842 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981868 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc9gm\" (UniqueName: \"kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981910 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.981952 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd\") pod \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\" (UID: \"7ea5a414-ff51-4622-8c07-4e1bf562d4c8\") " Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.984210 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.984272 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.987870 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm" (OuterVolumeSpecName: "kube-api-access-lc9gm") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "kube-api-access-lc9gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.991739 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts" (OuterVolumeSpecName: "scripts") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:03:59 crc kubenswrapper[4970]: I1124 14:03:59.998962 4970 scope.go:117] "RemoveContainer" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.001292 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": container with ID starting with 501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619 not found: ID does not exist" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.001332 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} err="failed to get container status \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": rpc error: code = NotFound desc = could not find container \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": container with ID starting with 501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.001380 4970 scope.go:117] "RemoveContainer" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.001650 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": container with ID starting with dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448 not found: ID does not exist" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.001698 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} err="failed to get container status \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": rpc error: code = NotFound desc = could not find container \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": container with ID starting with dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.001722 4970 scope.go:117] "RemoveContainer" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.001956 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": container with ID starting with 19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298 not found: ID does not exist" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002108 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} err="failed to get container status \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": rpc error: code = NotFound desc = could not find container \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": container with ID starting with 19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002132 4970 scope.go:117] "RemoveContainer" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.002365 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": container with ID starting with 4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445 not found: ID does not exist" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002409 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} err="failed to get container status \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": rpc error: code = NotFound desc = could not find container \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": container with ID starting with 4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002451 4970 scope.go:117] "RemoveContainer" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002704 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} err="failed to get container status \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": rpc error: code = NotFound desc = could not find container \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": container with ID starting with 501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002746 4970 scope.go:117] "RemoveContainer" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002950 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} err="failed to get container status \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": rpc error: code = NotFound desc = could not find container \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": container with ID starting with dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.002965 4970 scope.go:117] "RemoveContainer" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.003339 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} err="failed to get container status \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": rpc error: code = NotFound desc = could not find container \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": container with ID starting with 19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.003384 4970 scope.go:117] "RemoveContainer" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.003767 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} err="failed to get container status \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": rpc error: code = NotFound desc = could not find container \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": container with ID starting with 4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.003810 4970 scope.go:117] "RemoveContainer" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.004119 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} err="failed to get container status \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": rpc error: code = NotFound desc = could not find container \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": container with ID starting with 501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.004165 4970 scope.go:117] "RemoveContainer" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.005841 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} err="failed to get container status \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": rpc error: code = NotFound desc = could not find container \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": container with ID starting with dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.005891 4970 scope.go:117] "RemoveContainer" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.006897 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} err="failed to get container status \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": rpc error: code = NotFound desc = could not find container \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": container with ID starting with 19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.006922 4970 scope.go:117] "RemoveContainer" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.009154 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} err="failed to get container status \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": rpc error: code = NotFound desc = could not find container \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": container with ID starting with 4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.009209 4970 scope.go:117] "RemoveContainer" containerID="501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.013494 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619"} err="failed to get container status \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": rpc error: code = NotFound desc = could not find container \"501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619\": container with ID starting with 501586cf4a59d1168e7e53d83dd8d58f81c52642ce78a8f57915affb7b6ea619 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.013533 4970 scope.go:117] "RemoveContainer" containerID="dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.016859 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.018739 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448"} err="failed to get container status \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": rpc error: code = NotFound desc = could not find container \"dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448\": container with ID starting with dda523f389db0f76005c4cd2db5f6f70ad7274544e737fd17971d871e4560448 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.018795 4970 scope.go:117] "RemoveContainer" containerID="19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.022737 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298"} err="failed to get container status \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": rpc error: code = NotFound desc = could not find container \"19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298\": container with ID starting with 19de36dde4d8ba492f58d8bbeebf6871344d7fd45c810bebc46f0d76293bf298 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.022791 4970 scope.go:117] "RemoveContainer" containerID="4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.026696 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445"} err="failed to get container status \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": rpc error: code = NotFound desc = could not find container \"4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445\": container with ID starting with 4226571679acbd824e407a27990cc04a35cd4390b041a7837c94e20803c86445 not found: ID does not exist" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.087419 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.087663 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc9gm\" (UniqueName: \"kubernetes.io/projected/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-kube-api-access-lc9gm\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.087756 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.087831 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.087895 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.113806 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.127813 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data" (OuterVolumeSpecName: "config-data") pod "7ea5a414-ff51-4622-8c07-4e1bf562d4c8" (UID: "7ea5a414-ff51-4622-8c07-4e1bf562d4c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.190134 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.190160 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ea5a414-ff51-4622-8c07-4e1bf562d4c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.214817 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.223689 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.256987 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.259287 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="proxy-httpd" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.259319 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="proxy-httpd" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.259334 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-central-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.259343 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-central-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.259380 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="sg-core" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.259390 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="sg-core" Nov 24 14:04:00 crc kubenswrapper[4970]: E1124 14:04:00.259420 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-notification-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.259433 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-notification-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.260588 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="proxy-httpd" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.260611 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-central-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.260634 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="sg-core" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.260647 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" containerName="ceilometer-notification-agent" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.280071 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.293280 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.317961 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.323067 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419639 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbcvp\" (UniqueName: \"kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419701 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419731 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419854 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419906 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.419929 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.420075 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.522590 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.522802 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbcvp\" (UniqueName: \"kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.522844 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.522869 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.522983 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.523048 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.523080 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.523637 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.523693 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.528107 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.528446 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.530317 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.539021 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.549171 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbcvp\" (UniqueName: \"kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp\") pod \"ceilometer-0\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.627421 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.961848 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:04:00 crc kubenswrapper[4970]: I1124 14:04:00.962180 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.022117 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.125380 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.481107 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea5a414-ff51-4622-8c07-4e1bf562d4c8" path="/var/lib/kubelet/pods/7ea5a414-ff51-4622-8c07-4e1bf562d4c8/volumes" Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.883393 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.936086 4970 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.936429 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerStarted","Data":"566e1e6cec36bb0dd3735d761141562b19665b035dee437249b20a4502391962"} Nov 24 14:04:01 crc kubenswrapper[4970]: I1124 14:04:01.937411 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 24 14:04:02 crc kubenswrapper[4970]: I1124 14:04:02.945872 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerStarted","Data":"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820"} Nov 24 14:04:03 crc kubenswrapper[4970]: I1124 14:04:03.958059 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerStarted","Data":"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa"} Nov 24 14:04:03 crc kubenswrapper[4970]: I1124 14:04:03.958531 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerStarted","Data":"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f"} Nov 24 14:04:05 crc kubenswrapper[4970]: I1124 14:04:05.988981 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerStarted","Data":"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c"} Nov 24 14:04:05 crc kubenswrapper[4970]: I1124 14:04:05.989325 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:04:06 crc kubenswrapper[4970]: I1124 14:04:06.026147 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.311687865 podStartE2EDuration="6.025903929s" podCreationTimestamp="2025-11-24 14:04:00 +0000 UTC" firstStartedPulling="2025-11-24 14:04:01.137961113 +0000 UTC m=+1056.425718406" lastFinishedPulling="2025-11-24 14:04:04.852177167 +0000 UTC m=+1060.139934470" observedRunningTime="2025-11-24 14:04:06.009879944 +0000 UTC m=+1061.297637277" watchObservedRunningTime="2025-11-24 14:04:06.025903929 +0000 UTC m=+1061.313661252" Nov 24 14:04:10 crc kubenswrapper[4970]: I1124 14:04:10.001203 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="9cfd8b77-9869-4672-865e-e39e8a752567" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.165:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:04:11 crc kubenswrapper[4970]: I1124 14:04:11.204324 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:04:11 crc kubenswrapper[4970]: I1124 14:04:11.204424 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:04:15 crc kubenswrapper[4970]: I1124 14:04:15.070850 4970 generic.go:334] "Generic (PLEG): container finished" podID="a45bf61d-f620-4455-b4b0-6d0973225fdb" containerID="7c43ed5de706ff312a95ad504212c825b98a056620129c5b1ae8b01e5d1d16b9" exitCode=0 Nov 24 14:04:15 crc kubenswrapper[4970]: I1124 14:04:15.070936 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" event={"ID":"a45bf61d-f620-4455-b4b0-6d0973225fdb","Type":"ContainerDied","Data":"7c43ed5de706ff312a95ad504212c825b98a056620129c5b1ae8b01e5d1d16b9"} Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.446182 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.523278 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.523528 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-central-agent" containerID="cri-o://84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820" gracePeriod=30 Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.525309 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="proxy-httpd" containerID="cri-o://c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c" gracePeriod=30 Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.525320 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-notification-agent" containerID="cri-o://d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f" gracePeriod=30 Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.525401 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="sg-core" containerID="cri-o://26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa" gracePeriod=30 Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.529930 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.552524 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle\") pod \"a45bf61d-f620-4455-b4b0-6d0973225fdb\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.552610 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts\") pod \"a45bf61d-f620-4455-b4b0-6d0973225fdb\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.552719 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jcjs\" (UniqueName: \"kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs\") pod \"a45bf61d-f620-4455-b4b0-6d0973225fdb\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.552848 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data\") pod \"a45bf61d-f620-4455-b4b0-6d0973225fdb\" (UID: \"a45bf61d-f620-4455-b4b0-6d0973225fdb\") " Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.558257 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts" (OuterVolumeSpecName: "scripts") pod "a45bf61d-f620-4455-b4b0-6d0973225fdb" (UID: "a45bf61d-f620-4455-b4b0-6d0973225fdb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.568353 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs" (OuterVolumeSpecName: "kube-api-access-8jcjs") pod "a45bf61d-f620-4455-b4b0-6d0973225fdb" (UID: "a45bf61d-f620-4455-b4b0-6d0973225fdb"). InnerVolumeSpecName "kube-api-access-8jcjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.604431 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data" (OuterVolumeSpecName: "config-data") pod "a45bf61d-f620-4455-b4b0-6d0973225fdb" (UID: "a45bf61d-f620-4455-b4b0-6d0973225fdb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.622363 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a45bf61d-f620-4455-b4b0-6d0973225fdb" (UID: "a45bf61d-f620-4455-b4b0-6d0973225fdb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.654975 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.655012 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.655024 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a45bf61d-f620-4455-b4b0-6d0973225fdb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:16 crc kubenswrapper[4970]: I1124 14:04:16.655034 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jcjs\" (UniqueName: \"kubernetes.io/projected/a45bf61d-f620-4455-b4b0-6d0973225fdb-kube-api-access-8jcjs\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.095817 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerID="c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c" exitCode=0 Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.096300 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerID="26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa" exitCode=2 Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.095902 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerDied","Data":"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c"} Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.096408 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerDied","Data":"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa"} Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.096496 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerID="84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820" exitCode=0 Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.096567 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerDied","Data":"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820"} Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.099057 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" event={"ID":"a45bf61d-f620-4455-b4b0-6d0973225fdb","Type":"ContainerDied","Data":"fa9917de1e922225da44da3e12c956f1e3737be1522f80aabeb03841ed149866"} Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.099144 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa9917de1e922225da44da3e12c956f1e3737be1522f80aabeb03841ed149866" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.099162 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7mjd6" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.190840 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:04:17 crc kubenswrapper[4970]: E1124 14:04:17.191287 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a45bf61d-f620-4455-b4b0-6d0973225fdb" containerName="nova-cell0-conductor-db-sync" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.191307 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a45bf61d-f620-4455-b4b0-6d0973225fdb" containerName="nova-cell0-conductor-db-sync" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.191533 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a45bf61d-f620-4455-b4b0-6d0973225fdb" containerName="nova-cell0-conductor-db-sync" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.192369 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.196502 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jt59z" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.196815 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.203206 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.368501 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.368623 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64z7w\" (UniqueName: \"kubernetes.io/projected/e8645891-3d8f-4506-8139-d13f7ee73054-kube-api-access-64z7w\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.368842 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.470430 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.470540 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.470615 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64z7w\" (UniqueName: \"kubernetes.io/projected/e8645891-3d8f-4506-8139-d13f7ee73054-kube-api-access-64z7w\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.477267 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.478138 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8645891-3d8f-4506-8139-d13f7ee73054-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.510678 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64z7w\" (UniqueName: \"kubernetes.io/projected/e8645891-3d8f-4506-8139-d13f7ee73054-kube-api-access-64z7w\") pod \"nova-cell0-conductor-0\" (UID: \"e8645891-3d8f-4506-8139-d13f7ee73054\") " pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.511193 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:17 crc kubenswrapper[4970]: I1124 14:04:17.977562 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 14:04:18 crc kubenswrapper[4970]: I1124 14:04:18.110323 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e8645891-3d8f-4506-8139-d13f7ee73054","Type":"ContainerStarted","Data":"3b5fe128efd8082089358f49a7bf07f5341ef9650aefc0023781ffbf5994a447"} Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.122954 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e8645891-3d8f-4506-8139-d13f7ee73054","Type":"ContainerStarted","Data":"e76b3f0710784189da513a9b2b2bcdf8f508f5764c91dd7918670f4b0e5b0d9e"} Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.123294 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.141249 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.141225559 podStartE2EDuration="2.141225559s" podCreationTimestamp="2025-11-24 14:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:19.138499487 +0000 UTC m=+1074.426256810" watchObservedRunningTime="2025-11-24 14:04:19.141225559 +0000 UTC m=+1074.428982852" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.798090 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.921895 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.921955 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922061 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922136 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922169 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922240 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922268 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbcvp\" (UniqueName: \"kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp\") pod \"e0d8a43c-ef87-4c12-ba3b-365d96015853\" (UID: \"e0d8a43c-ef87-4c12-ba3b-365d96015853\") " Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922512 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922656 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.922796 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.933738 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts" (OuterVolumeSpecName: "scripts") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.934052 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp" (OuterVolumeSpecName: "kube-api-access-fbcvp") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "kube-api-access-fbcvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:19 crc kubenswrapper[4970]: I1124 14:04:19.965775 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.003431 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.024101 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.024137 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbcvp\" (UniqueName: \"kubernetes.io/projected/e0d8a43c-ef87-4c12-ba3b-365d96015853-kube-api-access-fbcvp\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.024147 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.024155 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.024164 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e0d8a43c-ef87-4c12-ba3b-365d96015853-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.027275 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data" (OuterVolumeSpecName: "config-data") pod "e0d8a43c-ef87-4c12-ba3b-365d96015853" (UID: "e0d8a43c-ef87-4c12-ba3b-365d96015853"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.125841 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0d8a43c-ef87-4c12-ba3b-365d96015853-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.137267 4970 generic.go:334] "Generic (PLEG): container finished" podID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerID="d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f" exitCode=0 Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.137351 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.137347 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerDied","Data":"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f"} Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.137423 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e0d8a43c-ef87-4c12-ba3b-365d96015853","Type":"ContainerDied","Data":"566e1e6cec36bb0dd3735d761141562b19665b035dee437249b20a4502391962"} Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.137454 4970 scope.go:117] "RemoveContainer" containerID="c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.178845 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.185454 4970 scope.go:117] "RemoveContainer" containerID="26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.196206 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.207090 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.207766 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="proxy-httpd" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.207784 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="proxy-httpd" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.207839 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-notification-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.207848 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-notification-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.207855 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="sg-core" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.207861 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="sg-core" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.207896 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-central-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.207904 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-central-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.208103 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-notification-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.208126 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="ceilometer-central-agent" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.208141 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="sg-core" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.208149 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" containerName="proxy-httpd" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.209951 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.215795 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.215903 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.223719 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.254474 4970 scope.go:117] "RemoveContainer" containerID="d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.283096 4970 scope.go:117] "RemoveContainer" containerID="84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.305382 4970 scope.go:117] "RemoveContainer" containerID="c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.306001 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c\": container with ID starting with c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c not found: ID does not exist" containerID="c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306049 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c"} err="failed to get container status \"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c\": rpc error: code = NotFound desc = could not find container \"c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c\": container with ID starting with c3c727f7c72be30e9223a31333d656d9c0caf5f7b52ffd35777ab96e47770d2c not found: ID does not exist" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306077 4970 scope.go:117] "RemoveContainer" containerID="26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.306557 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa\": container with ID starting with 26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa not found: ID does not exist" containerID="26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306579 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa"} err="failed to get container status \"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa\": rpc error: code = NotFound desc = could not find container \"26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa\": container with ID starting with 26caba4c16d99b306af290b973aa3e8d5a8571e8bdaf8f288aa1a4e04ef26efa not found: ID does not exist" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306602 4970 scope.go:117] "RemoveContainer" containerID="d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.306820 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f\": container with ID starting with d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f not found: ID does not exist" containerID="d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306839 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f"} err="failed to get container status \"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f\": rpc error: code = NotFound desc = could not find container \"d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f\": container with ID starting with d0e012f311e9f6e9e3c4bb9fc9c8b9abf6c4d1051b62f129307515d842e7d83f not found: ID does not exist" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.306851 4970 scope.go:117] "RemoveContainer" containerID="84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820" Nov 24 14:04:20 crc kubenswrapper[4970]: E1124 14:04:20.307032 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820\": container with ID starting with 84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820 not found: ID does not exist" containerID="84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.307066 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820"} err="failed to get container status \"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820\": rpc error: code = NotFound desc = could not find container \"84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820\": container with ID starting with 84b79a04d32b335033162cd9bfda482b0bd0741dbb74834c85348468c9142820 not found: ID does not exist" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.329437 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmkhv\" (UniqueName: \"kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.329485 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.329510 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.329883 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.330044 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.330429 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.330515 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.432281 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmkhv\" (UniqueName: \"kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.432339 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.432359 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433094 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433172 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433200 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433207 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433284 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.433536 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.437080 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.437668 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.441843 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.445175 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.463517 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmkhv\" (UniqueName: \"kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv\") pod \"ceilometer-0\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " pod="openstack/ceilometer-0" Nov 24 14:04:20 crc kubenswrapper[4970]: I1124 14:04:20.569149 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:21 crc kubenswrapper[4970]: I1124 14:04:21.032818 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:21 crc kubenswrapper[4970]: W1124 14:04:21.042130 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90d420e_5ea1_46f4_8b51_ba8b1957143b.slice/crio-8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad WatchSource:0}: Error finding container 8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad: Status 404 returned error can't find the container with id 8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad Nov 24 14:04:21 crc kubenswrapper[4970]: I1124 14:04:21.148209 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerStarted","Data":"8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad"} Nov 24 14:04:21 crc kubenswrapper[4970]: I1124 14:04:21.480923 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0d8a43c-ef87-4c12-ba3b-365d96015853" path="/var/lib/kubelet/pods/e0d8a43c-ef87-4c12-ba3b-365d96015853/volumes" Nov 24 14:04:21 crc kubenswrapper[4970]: I1124 14:04:21.958306 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:21 crc kubenswrapper[4970]: I1124 14:04:21.958791 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" containerName="kube-state-metrics" containerID="cri-o://d8eab8f64df417c0162763f8b1b084f62121581fc02b1f4a709d3d953f7a7ce8" gracePeriod=30 Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.165411 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerStarted","Data":"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77"} Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.168047 4970 generic.go:334] "Generic (PLEG): container finished" podID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" containerID="d8eab8f64df417c0162763f8b1b084f62121581fc02b1f4a709d3d953f7a7ce8" exitCode=2 Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.168107 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26f07678-0be1-4b4d-9754-8a8c54b8d82a","Type":"ContainerDied","Data":"d8eab8f64df417c0162763f8b1b084f62121581fc02b1f4a709d3d953f7a7ce8"} Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.427970 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.567274 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpxpw\" (UniqueName: \"kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw\") pod \"26f07678-0be1-4b4d-9754-8a8c54b8d82a\" (UID: \"26f07678-0be1-4b4d-9754-8a8c54b8d82a\") " Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.572495 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw" (OuterVolumeSpecName: "kube-api-access-kpxpw") pod "26f07678-0be1-4b4d-9754-8a8c54b8d82a" (UID: "26f07678-0be1-4b4d-9754-8a8c54b8d82a"). InnerVolumeSpecName "kube-api-access-kpxpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:22 crc kubenswrapper[4970]: I1124 14:04:22.669047 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpxpw\" (UniqueName: \"kubernetes.io/projected/26f07678-0be1-4b4d-9754-8a8c54b8d82a-kube-api-access-kpxpw\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.178708 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerStarted","Data":"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6"} Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.179972 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26f07678-0be1-4b4d-9754-8a8c54b8d82a","Type":"ContainerDied","Data":"ebff66a78c61b7d9033174f06a5f6032ca7f27a723d91064af49be3b55b1766a"} Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.180018 4970 scope.go:117] "RemoveContainer" containerID="d8eab8f64df417c0162763f8b1b084f62121581fc02b1f4a709d3d953f7a7ce8" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.180044 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.215758 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.225962 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.240209 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:23 crc kubenswrapper[4970]: E1124 14:04:23.240606 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" containerName="kube-state-metrics" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.240619 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" containerName="kube-state-metrics" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.240791 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" containerName="kube-state-metrics" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.241538 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.244887 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.248591 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.259004 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.385099 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.385408 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knjqj\" (UniqueName: \"kubernetes.io/projected/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-api-access-knjqj\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.385517 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.385600 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.481187 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f07678-0be1-4b4d-9754-8a8c54b8d82a" path="/var/lib/kubelet/pods/26f07678-0be1-4b4d-9754-8a8c54b8d82a/volumes" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.486969 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knjqj\" (UniqueName: \"kubernetes.io/projected/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-api-access-knjqj\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.487250 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.487278 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.487312 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.493703 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.493878 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.493879 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.504708 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knjqj\" (UniqueName: \"kubernetes.io/projected/3753fd35-b151-4bde-aa8c-a9c5f0a65f3f-kube-api-access-knjqj\") pod \"kube-state-metrics-0\" (UID: \"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f\") " pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.560343 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 14:04:23 crc kubenswrapper[4970]: I1124 14:04:23.876235 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:24 crc kubenswrapper[4970]: I1124 14:04:24.062013 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 14:04:24 crc kubenswrapper[4970]: I1124 14:04:24.187602 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f","Type":"ContainerStarted","Data":"018be6574bc74efb2e8c04526709490fac74872ad7cfe296225b54f7f9ea1611"} Nov 24 14:04:24 crc kubenswrapper[4970]: I1124 14:04:24.194488 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerStarted","Data":"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453"} Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.210353 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerStarted","Data":"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3"} Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.211026 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.210715 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-central-agent" containerID="cri-o://4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77" gracePeriod=30 Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.211171 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="proxy-httpd" containerID="cri-o://eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3" gracePeriod=30 Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.211417 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-notification-agent" containerID="cri-o://c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6" gracePeriod=30 Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.211509 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="sg-core" containerID="cri-o://6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453" gracePeriod=30 Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.212619 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3753fd35-b151-4bde-aa8c-a9c5f0a65f3f","Type":"ContainerStarted","Data":"2806a408b8a8eea5179dede841c715c71b9054c5360cd9bf32e66a074c9da50c"} Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.212780 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.241049 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.335931714 podStartE2EDuration="5.241032077s" podCreationTimestamp="2025-11-24 14:04:20 +0000 UTC" firstStartedPulling="2025-11-24 14:04:21.044963301 +0000 UTC m=+1076.332720594" lastFinishedPulling="2025-11-24 14:04:24.950063664 +0000 UTC m=+1080.237820957" observedRunningTime="2025-11-24 14:04:25.237519605 +0000 UTC m=+1080.525276908" watchObservedRunningTime="2025-11-24 14:04:25.241032077 +0000 UTC m=+1080.528789370" Nov 24 14:04:25 crc kubenswrapper[4970]: I1124 14:04:25.258142 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.885292178 podStartE2EDuration="2.25812092s" podCreationTimestamp="2025-11-24 14:04:23 +0000 UTC" firstStartedPulling="2025-11-24 14:04:24.067195393 +0000 UTC m=+1079.354952686" lastFinishedPulling="2025-11-24 14:04:24.440024145 +0000 UTC m=+1079.727781428" observedRunningTime="2025-11-24 14:04:25.254000631 +0000 UTC m=+1080.541757924" watchObservedRunningTime="2025-11-24 14:04:25.25812092 +0000 UTC m=+1080.545878213" Nov 24 14:04:26 crc kubenswrapper[4970]: I1124 14:04:26.225707 4970 generic.go:334] "Generic (PLEG): container finished" podID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerID="6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453" exitCode=2 Nov 24 14:04:26 crc kubenswrapper[4970]: I1124 14:04:26.226004 4970 generic.go:334] "Generic (PLEG): container finished" podID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerID="c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6" exitCode=0 Nov 24 14:04:26 crc kubenswrapper[4970]: I1124 14:04:26.225794 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerDied","Data":"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453"} Nov 24 14:04:26 crc kubenswrapper[4970]: I1124 14:04:26.226909 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerDied","Data":"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6"} Nov 24 14:04:27 crc kubenswrapper[4970]: I1124 14:04:27.559624 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.013665 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bqxgx"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.014913 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.018223 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.018232 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.026089 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bqxgx"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.153703 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.155178 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.160353 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.171621 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.174164 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.174224 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.174310 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l6sz\" (UniqueName: \"kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.174338 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.246567 4970 generic.go:334] "Generic (PLEG): container finished" podID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerID="4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77" exitCode=0 Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.246618 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerDied","Data":"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77"} Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276144 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276218 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276278 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8jpr\" (UniqueName: \"kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276369 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l6sz\" (UniqueName: \"kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276411 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276441 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.276475 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.285707 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.286481 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.307857 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l6sz\" (UniqueName: \"kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.311287 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data\") pod \"nova-cell0-cell-mapping-bqxgx\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.333073 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.340911 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.342430 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.356091 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.384702 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.384763 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.384889 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8jpr\" (UniqueName: \"kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.396004 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.396483 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.403908 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.405303 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.412058 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.428538 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8jpr\" (UniqueName: \"kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr\") pod \"nova-scheduler-0\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.433667 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.449550 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.479062 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.500905 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5gh\" (UniqueName: \"kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.500972 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.501091 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.501114 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.599476 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.606673 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611084 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611143 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611242 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp9qg\" (UniqueName: \"kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611331 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611358 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611411 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5gh\" (UniqueName: \"kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.611429 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.612119 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.619437 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.619655 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.621071 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.624488 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.625439 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.639682 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.650974 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.671066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5gh\" (UniqueName: \"kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh\") pod \"nova-metadata-0\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713731 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713768 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713795 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713829 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp9qg\" (UniqueName: \"kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713850 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713885 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713914 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713957 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713977 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.713995 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.714023 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l52p\" (UniqueName: \"kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.714055 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b67nx\" (UniqueName: \"kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.714085 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.718419 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.735292 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.750386 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp9qg\" (UniqueName: \"kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg\") pod \"nova-cell1-novncproxy-0\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816029 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816091 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816114 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816150 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l52p\" (UniqueName: \"kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816178 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b67nx\" (UniqueName: \"kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816227 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816242 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816261 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816295 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.816327 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.817082 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.817626 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.818119 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.818332 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.818945 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.819439 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.824198 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.824205 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.836177 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.841938 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.852151 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l52p\" (UniqueName: \"kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p\") pod \"nova-api-0\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.854129 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b67nx\" (UniqueName: \"kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx\") pod \"dnsmasq-dns-64dbf5859c-2k4gm\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.945758 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:28 crc kubenswrapper[4970]: I1124 14:04:28.974995 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.101601 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bqxgx"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.193303 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.296809 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ec89865-98e3-46d3-84cc-b4e5120b992f","Type":"ContainerStarted","Data":"fb9c85057444f17756ff32ec01ff1bae6723c326f9992bef7ff71bbe7e592b16"} Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.298840 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bqxgx" event={"ID":"c96d3bcf-7d2f-4946-a998-2567e91b275c","Type":"ContainerStarted","Data":"c84705250b890dc30563dcc3805ac60fecb1269e12856a206fa4efbed104d75c"} Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.349947 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjs98"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.351595 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.354828 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.355351 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.361505 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjs98"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.413049 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.426952 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.427008 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mll\" (UniqueName: \"kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.427031 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.427116 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.528867 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.529013 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.529066 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mll\" (UniqueName: \"kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.529100 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.535180 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.536935 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.554729 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mll\" (UniqueName: \"kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.555498 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data\") pod \"nova-cell1-conductor-db-sync-gjs98\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.586745 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.596490 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.684452 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:29 crc kubenswrapper[4970]: I1124 14:04:29.727505 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.145311 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjs98"] Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.311269 4970 generic.go:334] "Generic (PLEG): container finished" podID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerID="485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e" exitCode=0 Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.311451 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" event={"ID":"d498d983-283c-4e6d-a2bb-0b143562a2c1","Type":"ContainerDied","Data":"485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.311540 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" event={"ID":"d498d983-283c-4e6d-a2bb-0b143562a2c1","Type":"ContainerStarted","Data":"c63bb66c05e0ff3d079f9997282b4e9153c74b093f27389a3427adf1455ce4fd"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.314423 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"26a9ee71-436d-4a52-9685-ab16d64796d1","Type":"ContainerStarted","Data":"0c4364bb6a0ebfd6e25cbfa233bf9781afe929e32b8ac96443c5f0e92fc55e3d"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.316937 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerStarted","Data":"624fd0a3f3c4f0db5ba1a1199ee412463552a6d8a7102d7ce431d9ec4bb4a853"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.319687 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerStarted","Data":"06820e82571649efd821f88ca43bde4507e5f7eec90d3c2509d9951c865463cd"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.321599 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjs98" event={"ID":"478cab1a-9f02-4f21-bae0-5f11ddc62665","Type":"ContainerStarted","Data":"755a000d4f23852e7326fcfcd44db7aef522dc795ee98e97ef4ae4b9293320ec"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.324340 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bqxgx" event={"ID":"c96d3bcf-7d2f-4946-a998-2567e91b275c","Type":"ContainerStarted","Data":"11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825"} Nov 24 14:04:30 crc kubenswrapper[4970]: I1124 14:04:30.355144 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bqxgx" podStartSLOduration=3.355121198 podStartE2EDuration="3.355121198s" podCreationTimestamp="2025-11-24 14:04:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:30.347955218 +0000 UTC m=+1085.635712511" watchObservedRunningTime="2025-11-24 14:04:30.355121198 +0000 UTC m=+1085.642878491" Nov 24 14:04:31 crc kubenswrapper[4970]: I1124 14:04:31.344049 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjs98" event={"ID":"478cab1a-9f02-4f21-bae0-5f11ddc62665","Type":"ContainerStarted","Data":"040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942"} Nov 24 14:04:31 crc kubenswrapper[4970]: I1124 14:04:31.346565 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" event={"ID":"d498d983-283c-4e6d-a2bb-0b143562a2c1","Type":"ContainerStarted","Data":"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584"} Nov 24 14:04:31 crc kubenswrapper[4970]: I1124 14:04:31.365087 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-gjs98" podStartSLOduration=2.365072398 podStartE2EDuration="2.365072398s" podCreationTimestamp="2025-11-24 14:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:31.357482238 +0000 UTC m=+1086.645239531" watchObservedRunningTime="2025-11-24 14:04:31.365072398 +0000 UTC m=+1086.652829691" Nov 24 14:04:31 crc kubenswrapper[4970]: I1124 14:04:31.390758 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" podStartSLOduration=3.390737918 podStartE2EDuration="3.390737918s" podCreationTimestamp="2025-11-24 14:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:31.385989333 +0000 UTC m=+1086.673746636" watchObservedRunningTime="2025-11-24 14:04:31.390737918 +0000 UTC m=+1086.678495211" Nov 24 14:04:32 crc kubenswrapper[4970]: I1124 14:04:32.064458 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:04:32 crc kubenswrapper[4970]: I1124 14:04:32.076306 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:04:32 crc kubenswrapper[4970]: I1124 14:04:32.358253 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.368250 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"26a9ee71-436d-4a52-9685-ab16d64796d1","Type":"ContainerStarted","Data":"26e36d6d8d834f11c4fc7136e5c9bb504ac9e0413b4ead7d36b8eb75e0a031fd"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.368594 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="26a9ee71-436d-4a52-9685-ab16d64796d1" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://26e36d6d8d834f11c4fc7136e5c9bb504ac9e0413b4ead7d36b8eb75e0a031fd" gracePeriod=30 Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.371790 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerStarted","Data":"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.371830 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerStarted","Data":"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.371963 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-log" containerID="cri-o://64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef" gracePeriod=30 Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.372093 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-metadata" containerID="cri-o://74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b" gracePeriod=30 Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.375161 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerStarted","Data":"130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.375222 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerStarted","Data":"418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.384301 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ec89865-98e3-46d3-84cc-b4e5120b992f","Type":"ContainerStarted","Data":"65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb"} Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.413109 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.188307266 podStartE2EDuration="5.413088956s" podCreationTimestamp="2025-11-24 14:04:28 +0000 UTC" firstStartedPulling="2025-11-24 14:04:29.420501093 +0000 UTC m=+1084.708258386" lastFinishedPulling="2025-11-24 14:04:32.645282783 +0000 UTC m=+1087.933040076" observedRunningTime="2025-11-24 14:04:33.407358074 +0000 UTC m=+1088.695115377" watchObservedRunningTime="2025-11-24 14:04:33.413088956 +0000 UTC m=+1088.700846249" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.435731 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.006981949 podStartE2EDuration="5.435706485s" podCreationTimestamp="2025-11-24 14:04:28 +0000 UTC" firstStartedPulling="2025-11-24 14:04:29.215640023 +0000 UTC m=+1084.503397316" lastFinishedPulling="2025-11-24 14:04:32.644364559 +0000 UTC m=+1087.932121852" observedRunningTime="2025-11-24 14:04:33.42645155 +0000 UTC m=+1088.714208843" watchObservedRunningTime="2025-11-24 14:04:33.435706485 +0000 UTC m=+1088.723463788" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.449069 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.401678551 podStartE2EDuration="5.449043848s" podCreationTimestamp="2025-11-24 14:04:28 +0000 UTC" firstStartedPulling="2025-11-24 14:04:29.595831691 +0000 UTC m=+1084.883588984" lastFinishedPulling="2025-11-24 14:04:32.643196988 +0000 UTC m=+1087.930954281" observedRunningTime="2025-11-24 14:04:33.439663 +0000 UTC m=+1088.727420293" watchObservedRunningTime="2025-11-24 14:04:33.449043848 +0000 UTC m=+1088.736801141" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.463061 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.401223079 podStartE2EDuration="5.463043989s" podCreationTimestamp="2025-11-24 14:04:28 +0000 UTC" firstStartedPulling="2025-11-24 14:04:29.606690579 +0000 UTC m=+1084.894447872" lastFinishedPulling="2025-11-24 14:04:32.668511489 +0000 UTC m=+1087.956268782" observedRunningTime="2025-11-24 14:04:33.459979808 +0000 UTC m=+1088.747737091" watchObservedRunningTime="2025-11-24 14:04:33.463043989 +0000 UTC m=+1088.750801282" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.487862 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.592440 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.837048 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.837357 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:04:33 crc kubenswrapper[4970]: I1124 14:04:33.843194 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:04:34 crc kubenswrapper[4970]: I1124 14:04:34.398535 4970 generic.go:334] "Generic (PLEG): container finished" podID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerID="64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef" exitCode=143 Nov 24 14:04:34 crc kubenswrapper[4970]: I1124 14:04:34.400284 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerDied","Data":"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef"} Nov 24 14:04:37 crc kubenswrapper[4970]: I1124 14:04:37.429900 4970 generic.go:334] "Generic (PLEG): container finished" podID="478cab1a-9f02-4f21-bae0-5f11ddc62665" containerID="040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942" exitCode=0 Nov 24 14:04:37 crc kubenswrapper[4970]: I1124 14:04:37.429972 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjs98" event={"ID":"478cab1a-9f02-4f21-bae0-5f11ddc62665","Type":"ContainerDied","Data":"040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942"} Nov 24 14:04:37 crc kubenswrapper[4970]: I1124 14:04:37.433767 4970 generic.go:334] "Generic (PLEG): container finished" podID="c96d3bcf-7d2f-4946-a998-2567e91b275c" containerID="11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825" exitCode=0 Nov 24 14:04:37 crc kubenswrapper[4970]: I1124 14:04:37.433805 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bqxgx" event={"ID":"c96d3bcf-7d2f-4946-a998-2567e91b275c","Type":"ContainerDied","Data":"11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825"} Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.480077 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.531564 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.938862 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.946961 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.947054 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:04:38 crc kubenswrapper[4970]: I1124 14:04:38.950028 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:38.985732 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029479 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle\") pod \"c96d3bcf-7d2f-4946-a998-2567e91b275c\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029520 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data\") pod \"c96d3bcf-7d2f-4946-a998-2567e91b275c\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029593 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l6sz\" (UniqueName: \"kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz\") pod \"c96d3bcf-7d2f-4946-a998-2567e91b275c\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029686 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle\") pod \"478cab1a-9f02-4f21-bae0-5f11ddc62665\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029770 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts\") pod \"478cab1a-9f02-4f21-bae0-5f11ddc62665\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029802 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mll\" (UniqueName: \"kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll\") pod \"478cab1a-9f02-4f21-bae0-5f11ddc62665\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029831 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data\") pod \"478cab1a-9f02-4f21-bae0-5f11ddc62665\" (UID: \"478cab1a-9f02-4f21-bae0-5f11ddc62665\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.029861 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts\") pod \"c96d3bcf-7d2f-4946-a998-2567e91b275c\" (UID: \"c96d3bcf-7d2f-4946-a998-2567e91b275c\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.037031 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll" (OuterVolumeSpecName: "kube-api-access-v4mll") pod "478cab1a-9f02-4f21-bae0-5f11ddc62665" (UID: "478cab1a-9f02-4f21-bae0-5f11ddc62665"). InnerVolumeSpecName "kube-api-access-v4mll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.045926 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz" (OuterVolumeSpecName: "kube-api-access-4l6sz") pod "c96d3bcf-7d2f-4946-a998-2567e91b275c" (UID: "c96d3bcf-7d2f-4946-a998-2567e91b275c"). InnerVolumeSpecName "kube-api-access-4l6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.052637 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts" (OuterVolumeSpecName: "scripts") pod "c96d3bcf-7d2f-4946-a998-2567e91b275c" (UID: "c96d3bcf-7d2f-4946-a998-2567e91b275c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.067867 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts" (OuterVolumeSpecName: "scripts") pod "478cab1a-9f02-4f21-bae0-5f11ddc62665" (UID: "478cab1a-9f02-4f21-bae0-5f11ddc62665"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.093113 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.093357 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="dnsmasq-dns" containerID="cri-o://c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f" gracePeriod=10 Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.104853 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c96d3bcf-7d2f-4946-a998-2567e91b275c" (UID: "c96d3bcf-7d2f-4946-a998-2567e91b275c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.130441 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "478cab1a-9f02-4f21-bae0-5f11ddc62665" (UID: "478cab1a-9f02-4f21-bae0-5f11ddc62665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132505 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132529 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mll\" (UniqueName: \"kubernetes.io/projected/478cab1a-9f02-4f21-bae0-5f11ddc62665-kube-api-access-v4mll\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132541 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132550 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132559 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l6sz\" (UniqueName: \"kubernetes.io/projected/c96d3bcf-7d2f-4946-a998-2567e91b275c-kube-api-access-4l6sz\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.132780 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.136075 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data" (OuterVolumeSpecName: "config-data") pod "478cab1a-9f02-4f21-bae0-5f11ddc62665" (UID: "478cab1a-9f02-4f21-bae0-5f11ddc62665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.191102 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data" (OuterVolumeSpecName: "config-data") pod "c96d3bcf-7d2f-4946-a998-2567e91b275c" (UID: "c96d3bcf-7d2f-4946-a998-2567e91b275c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.235153 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/478cab1a-9f02-4f21-bae0-5f11ddc62665-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.235506 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c96d3bcf-7d2f-4946-a998-2567e91b275c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.458951 4970 generic.go:334] "Generic (PLEG): container finished" podID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerID="c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f" exitCode=0 Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.459041 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" event={"ID":"eb87ce82-23c4-4610-a607-c648b7c88b81","Type":"ContainerDied","Data":"c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f"} Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.465554 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-gjs98" event={"ID":"478cab1a-9f02-4f21-bae0-5f11ddc62665","Type":"ContainerDied","Data":"755a000d4f23852e7326fcfcd44db7aef522dc795ee98e97ef4ae4b9293320ec"} Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.465681 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="755a000d4f23852e7326fcfcd44db7aef522dc795ee98e97ef4ae4b9293320ec" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.465761 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-gjs98" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.469300 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bqxgx" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.469358 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bqxgx" event={"ID":"c96d3bcf-7d2f-4946-a998-2567e91b275c","Type":"ContainerDied","Data":"c84705250b890dc30563dcc3805ac60fecb1269e12856a206fa4efbed104d75c"} Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.469393 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c84705250b890dc30563dcc3805ac60fecb1269e12856a206fa4efbed104d75c" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.562225 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:04:39 crc kubenswrapper[4970]: E1124 14:04:39.562645 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c96d3bcf-7d2f-4946-a998-2567e91b275c" containerName="nova-manage" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.562657 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c96d3bcf-7d2f-4946-a998-2567e91b275c" containerName="nova-manage" Nov 24 14:04:39 crc kubenswrapper[4970]: E1124 14:04:39.562682 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="478cab1a-9f02-4f21-bae0-5f11ddc62665" containerName="nova-cell1-conductor-db-sync" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.562688 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="478cab1a-9f02-4f21-bae0-5f11ddc62665" containerName="nova-cell1-conductor-db-sync" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.562876 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="478cab1a-9f02-4f21-bae0-5f11ddc62665" containerName="nova-cell1-conductor-db-sync" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.562895 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c96d3bcf-7d2f-4946-a998-2567e91b275c" containerName="nova-manage" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.563496 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.567089 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.569543 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.573543 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.622602 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.644337 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f6rb\" (UniqueName: \"kubernetes.io/projected/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-kube-api-access-6f6rb\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.644486 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.644537 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.745879 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.745952 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5gbd\" (UniqueName: \"kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746025 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746068 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746173 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746219 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb\") pod \"eb87ce82-23c4-4610-a607-c648b7c88b81\" (UID: \"eb87ce82-23c4-4610-a607-c648b7c88b81\") " Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746565 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746646 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.746743 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f6rb\" (UniqueName: \"kubernetes.io/projected/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-kube-api-access-6f6rb\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.753428 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.756290 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.766106 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd" (OuterVolumeSpecName: "kube-api-access-b5gbd") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "kube-api-access-b5gbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.775619 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f6rb\" (UniqueName: \"kubernetes.io/projected/0c35af6c-0331-4c98-9cb4-116c2ed45ec7-kube-api-access-6f6rb\") pod \"nova-cell1-conductor-0\" (UID: \"0c35af6c-0331-4c98-9cb4-116c2ed45ec7\") " pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.810388 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.816261 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.817743 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-log" containerID="cri-o://418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5" gracePeriod=30 Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.818366 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-api" containerID="cri-o://130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b" gracePeriod=30 Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.822030 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.823038 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": EOF" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.823117 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": EOF" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.844893 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config" (OuterVolumeSpecName: "config") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.848253 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.848276 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.848288 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5gbd\" (UniqueName: \"kubernetes.io/projected/eb87ce82-23c4-4610-a607-c648b7c88b81-kube-api-access-b5gbd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.848301 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.855039 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.864095 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb87ce82-23c4-4610-a607-c648b7c88b81" (UID: "eb87ce82-23c4-4610-a607-c648b7c88b81"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.903125 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.949651 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:39 crc kubenswrapper[4970]: I1124 14:04:39.949682 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb87ce82-23c4-4610-a607-c648b7c88b81-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.073768 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.372483 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.491166 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0c35af6c-0331-4c98-9cb4-116c2ed45ec7","Type":"ContainerStarted","Data":"89cd554e70a89497e577a77fae37611b9a88cde29220b9ae6fecf2698d92e130"} Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.500128 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" event={"ID":"eb87ce82-23c4-4610-a607-c648b7c88b81","Type":"ContainerDied","Data":"947b7a79252a07fc3896e23d26f0ef4cce595f8dc977c469fe7b37f714bbf543"} Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.500241 4970 scope.go:117] "RemoveContainer" containerID="c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f" Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.500305 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7965876c4f-fnzbv" Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.506178 4970 generic.go:334] "Generic (PLEG): container finished" podID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerID="418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5" exitCode=143 Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.506354 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerDied","Data":"418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5"} Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.525050 4970 scope.go:117] "RemoveContainer" containerID="a076460ac0bd5ec78eec01f4ef65fe4f5acb9f6053f7cfaea8d0553ea7d8d65a" Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.564763 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:04:40 crc kubenswrapper[4970]: I1124 14:04:40.571384 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7965876c4f-fnzbv"] Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.204900 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.205542 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.484465 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" path="/var/lib/kubelet/pods/eb87ce82-23c4-4610-a607-c648b7c88b81/volumes" Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.521791 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0c35af6c-0331-4c98-9cb4-116c2ed45ec7","Type":"ContainerStarted","Data":"556ce3dedb953cbdd39e41ddbfdf145cfe257167d62470f2dc6656d01346e4e0"} Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.521851 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.523022 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerName="nova-scheduler-scheduler" containerID="cri-o://65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" gracePeriod=30 Nov 24 14:04:41 crc kubenswrapper[4970]: I1124 14:04:41.543924 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.543901681 podStartE2EDuration="2.543901681s" podCreationTimestamp="2025-11-24 14:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:41.538939609 +0000 UTC m=+1096.826696912" watchObservedRunningTime="2025-11-24 14:04:41.543901681 +0000 UTC m=+1096.831658974" Nov 24 14:04:43 crc kubenswrapper[4970]: E1124 14:04:43.482098 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:04:43 crc kubenswrapper[4970]: E1124 14:04:43.484814 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:04:43 crc kubenswrapper[4970]: E1124 14:04:43.486934 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:04:43 crc kubenswrapper[4970]: E1124 14:04:43.487008 4970 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerName="nova-scheduler-scheduler" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.572187 4970 generic.go:334] "Generic (PLEG): container finished" podID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerID="65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" exitCode=0 Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.572277 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ec89865-98e3-46d3-84cc-b4e5120b992f","Type":"ContainerDied","Data":"65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb"} Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.711293 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.862245 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8jpr\" (UniqueName: \"kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr\") pod \"0ec89865-98e3-46d3-84cc-b4e5120b992f\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.862474 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle\") pod \"0ec89865-98e3-46d3-84cc-b4e5120b992f\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.862608 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data\") pod \"0ec89865-98e3-46d3-84cc-b4e5120b992f\" (UID: \"0ec89865-98e3-46d3-84cc-b4e5120b992f\") " Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.874259 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr" (OuterVolumeSpecName: "kube-api-access-m8jpr") pod "0ec89865-98e3-46d3-84cc-b4e5120b992f" (UID: "0ec89865-98e3-46d3-84cc-b4e5120b992f"). InnerVolumeSpecName "kube-api-access-m8jpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.900715 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data" (OuterVolumeSpecName: "config-data") pod "0ec89865-98e3-46d3-84cc-b4e5120b992f" (UID: "0ec89865-98e3-46d3-84cc-b4e5120b992f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.929819 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ec89865-98e3-46d3-84cc-b4e5120b992f" (UID: "0ec89865-98e3-46d3-84cc-b4e5120b992f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.966767 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.966816 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ec89865-98e3-46d3-84cc-b4e5120b992f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:44 crc kubenswrapper[4970]: I1124 14:04:44.966835 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8jpr\" (UniqueName: \"kubernetes.io/projected/0ec89865-98e3-46d3-84cc-b4e5120b992f-kube-api-access-m8jpr\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.584649 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.584823 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0ec89865-98e3-46d3-84cc-b4e5120b992f","Type":"ContainerDied","Data":"fb9c85057444f17756ff32ec01ff1bae6723c326f9992bef7ff71bbe7e592b16"} Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.585105 4970 scope.go:117] "RemoveContainer" containerID="65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.590188 4970 generic.go:334] "Generic (PLEG): container finished" podID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerID="130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b" exitCode=0 Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.590225 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerDied","Data":"130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b"} Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.608131 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.615744 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.641562 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:45 crc kubenswrapper[4970]: E1124 14:04:45.642011 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="init" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.642032 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="init" Nov 24 14:04:45 crc kubenswrapper[4970]: E1124 14:04:45.642065 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerName="nova-scheduler-scheduler" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.642074 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerName="nova-scheduler-scheduler" Nov 24 14:04:45 crc kubenswrapper[4970]: E1124 14:04:45.642099 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="dnsmasq-dns" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.642108 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="dnsmasq-dns" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.642334 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" containerName="nova-scheduler-scheduler" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.642369 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb87ce82-23c4-4610-a607-c648b7c88b81" containerName="dnsmasq-dns" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.643708 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.646331 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.654457 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.719072 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783231 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7l52p\" (UniqueName: \"kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p\") pod \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783274 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle\") pod \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783305 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs\") pod \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783403 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data\") pod \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\" (UID: \"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66\") " Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783719 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-656mn\" (UniqueName: \"kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783794 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.783832 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.784058 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs" (OuterVolumeSpecName: "logs") pod "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" (UID: "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.790413 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p" (OuterVolumeSpecName: "kube-api-access-7l52p") pod "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" (UID: "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66"). InnerVolumeSpecName "kube-api-access-7l52p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.816682 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" (UID: "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.826248 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data" (OuterVolumeSpecName: "config-data") pod "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" (UID: "93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.885837 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.886080 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.886557 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-656mn\" (UniqueName: \"kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.886771 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.886883 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7l52p\" (UniqueName: \"kubernetes.io/projected/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-kube-api-access-7l52p\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.887006 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.887150 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.889305 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.889702 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:45 crc kubenswrapper[4970]: I1124 14:04:45.907130 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-656mn\" (UniqueName: \"kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn\") pod \"nova-scheduler-0\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " pod="openstack/nova-scheduler-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.026503 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.494993 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.607381 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5","Type":"ContainerStarted","Data":"06bcb6ed4193114135b9f95b5d183fb77c38b8d374698f898b60658b4c04e3cc"} Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.613676 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66","Type":"ContainerDied","Data":"06820e82571649efd821f88ca43bde4507e5f7eec90d3c2509d9951c865463cd"} Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.613724 4970 scope.go:117] "RemoveContainer" containerID="130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.613825 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.653233 4970 scope.go:117] "RemoveContainer" containerID="418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.674791 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.682502 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.707500 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:46 crc kubenswrapper[4970]: E1124 14:04:46.708180 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-api" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.708273 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-api" Nov 24 14:04:46 crc kubenswrapper[4970]: E1124 14:04:46.708359 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-log" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.708412 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-log" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.708678 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-api" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.708776 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" containerName="nova-api-log" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.709871 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.711900 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.719165 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.801858 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.801926 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.802000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp8vz\" (UniqueName: \"kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.802058 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.903541 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp8vz\" (UniqueName: \"kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.903711 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.903795 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.903881 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.905185 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.909831 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.910481 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:46 crc kubenswrapper[4970]: I1124 14:04:46.935985 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp8vz\" (UniqueName: \"kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz\") pod \"nova-api-0\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " pod="openstack/nova-api-0" Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.038073 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.485278 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec89865-98e3-46d3-84cc-b4e5120b992f" path="/var/lib/kubelet/pods/0ec89865-98e3-46d3-84cc-b4e5120b992f/volumes" Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.486719 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66" path="/var/lib/kubelet/pods/93ab7a4e-1f4e-49f3-8cbc-343e2c8aef66/volumes" Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.623389 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.636423 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5","Type":"ContainerStarted","Data":"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946"} Nov 24 14:04:47 crc kubenswrapper[4970]: I1124 14:04:47.672469 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.672447243 podStartE2EDuration="2.672447243s" podCreationTimestamp="2025-11-24 14:04:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:47.658133463 +0000 UTC m=+1102.945890756" watchObservedRunningTime="2025-11-24 14:04:47.672447243 +0000 UTC m=+1102.960204546" Nov 24 14:04:48 crc kubenswrapper[4970]: I1124 14:04:48.651771 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerStarted","Data":"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c"} Nov 24 14:04:48 crc kubenswrapper[4970]: I1124 14:04:48.653046 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerStarted","Data":"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9"} Nov 24 14:04:48 crc kubenswrapper[4970]: I1124 14:04:48.653118 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerStarted","Data":"c8961bdf6451482004b799bb9056c29f4ebb66cf800ea765724cf1e9413bc426"} Nov 24 14:04:48 crc kubenswrapper[4970]: I1124 14:04:48.672410 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.672394308 podStartE2EDuration="2.672394308s" podCreationTimestamp="2025-11-24 14:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:04:48.6694344 +0000 UTC m=+1103.957191723" watchObservedRunningTime="2025-11-24 14:04:48.672394308 +0000 UTC m=+1103.960151601" Nov 24 14:04:49 crc kubenswrapper[4970]: I1124 14:04:49.952869 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 14:04:50 crc kubenswrapper[4970]: I1124 14:04:50.573779 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 14:04:51 crc kubenswrapper[4970]: I1124 14:04:51.027333 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.275884 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd498d983_283c_4e6d_a2bb_0b143562a2c1.slice/crio-conmon-485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd498d983_283c_4e6d_a2bb_0b143562a2c1.slice/crio-conmon-485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.276695 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd498d983_283c_4e6d_a2bb_0b143562a2c1.slice/crio-485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd498d983_283c_4e6d_a2bb_0b143562a2c1.slice/crio-485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.276741 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-755a000d4f23852e7326fcfcd44db7aef522dc795ee98e97ef4ae4b9293320ec": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-755a000d4f23852e7326fcfcd44db7aef522dc795ee98e97ef4ae4b9293320ec: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.276781 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-conmon-040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-conmon-040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.276809 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice/crio-040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277261 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-conmon-418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-conmon-418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277518 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice/crio-conmon-65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice/crio-conmon-65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277565 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bf37c2_7bb7_4d61_b856_b9550a89598d.slice/crio-conmon-64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bf37c2_7bb7_4d61_b856_b9550a89598d.slice/crio-conmon-64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277670 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice/crio-65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice/crio-65c09d0513270ee7d56df28e24c37c2d17e3a2309ce886058e7bcd3cd12224cb.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277709 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-418ccb5f6574cdbb3a8fa4a311f90046d8ffdee7679f9222c825e0b0cfb031d5.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277947 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bf37c2_7bb7_4d61_b856_b9550a89598d.slice/crio-64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41bf37c2_7bb7_4d61_b856_b9550a89598d.slice/crio-64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.277982 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-conmon-130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-conmon-130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: W1124 14:04:55.278188 4970 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-130261daf0aa21d64a23ac46efe3f787cd67cfc17e693ec683fd9b8d7514d38b.scope: no such file or directory Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.719019 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.752803 4970 generic.go:334] "Generic (PLEG): container finished" podID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerID="eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3" exitCode=137 Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.752869 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerDied","Data":"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3"} Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.752910 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.752968 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b90d420e-5ea1-46f4-8b51-ba8b1957143b","Type":"ContainerDied","Data":"8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad"} Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.753004 4970 scope.go:117] "RemoveContainer" containerID="eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.772607 4970 scope.go:117] "RemoveContainer" containerID="6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.798991 4970 scope.go:117] "RemoveContainer" containerID="c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806099 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806164 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806287 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806343 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806373 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmkhv\" (UniqueName: \"kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806390 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.806458 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data\") pod \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\" (UID: \"b90d420e-5ea1-46f4-8b51-ba8b1957143b\") " Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.807890 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.808368 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.813557 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts" (OuterVolumeSpecName: "scripts") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.816793 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv" (OuterVolumeSpecName: "kube-api-access-nmkhv") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "kube-api-access-nmkhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.832631 4970 scope.go:117] "RemoveContainer" containerID="4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.848054 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.908669 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.908697 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.908706 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmkhv\" (UniqueName: \"kubernetes.io/projected/b90d420e-5ea1-46f4-8b51-ba8b1957143b-kube-api-access-nmkhv\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.908714 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b90d420e-5ea1-46f4-8b51-ba8b1957143b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.908723 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.923637 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.929042 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data" (OuterVolumeSpecName: "config-data") pod "b90d420e-5ea1-46f4-8b51-ba8b1957143b" (UID: "b90d420e-5ea1-46f4-8b51-ba8b1957143b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.941154 4970 scope.go:117] "RemoveContainer" containerID="eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3" Nov 24 14:04:55 crc kubenswrapper[4970]: E1124 14:04:55.941853 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3\": container with ID starting with eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3 not found: ID does not exist" containerID="eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.941904 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3"} err="failed to get container status \"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3\": rpc error: code = NotFound desc = could not find container \"eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3\": container with ID starting with eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3 not found: ID does not exist" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.941932 4970 scope.go:117] "RemoveContainer" containerID="6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453" Nov 24 14:04:55 crc kubenswrapper[4970]: E1124 14:04:55.942542 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453\": container with ID starting with 6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453 not found: ID does not exist" containerID="6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.942607 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453"} err="failed to get container status \"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453\": rpc error: code = NotFound desc = could not find container \"6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453\": container with ID starting with 6c49db045c265e34bd6ea0d62bb52cb27d65b67b97f1f55cb64c926975665453 not found: ID does not exist" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.942642 4970 scope.go:117] "RemoveContainer" containerID="c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6" Nov 24 14:04:55 crc kubenswrapper[4970]: E1124 14:04:55.943044 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6\": container with ID starting with c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6 not found: ID does not exist" containerID="c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.943079 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6"} err="failed to get container status \"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6\": rpc error: code = NotFound desc = could not find container \"c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6\": container with ID starting with c4cf0c79d5dbd121c3a82b469a99f892953c86f1a89efd3f1d501f7e3ca2dfd6 not found: ID does not exist" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.943095 4970 scope.go:117] "RemoveContainer" containerID="4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77" Nov 24 14:04:55 crc kubenswrapper[4970]: E1124 14:04:55.943510 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77\": container with ID starting with 4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77 not found: ID does not exist" containerID="4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77" Nov 24 14:04:55 crc kubenswrapper[4970]: I1124 14:04:55.943537 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77"} err="failed to get container status \"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77\": rpc error: code = NotFound desc = could not find container \"4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77\": container with ID starting with 4ef10bf7b107c7ee5192d2635037150cf13e3f93665e914ad5968d81ce482d77 not found: ID does not exist" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.011159 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.011210 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b90d420e-5ea1-46f4-8b51-ba8b1957143b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.026869 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.076067 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.115095 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.150808 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.156933 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:56 crc kubenswrapper[4970]: E1124 14:04:56.157404 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-notification-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157421 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-notification-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: E1124 14:04:56.157437 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="proxy-httpd" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157446 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="proxy-httpd" Nov 24 14:04:56 crc kubenswrapper[4970]: E1124 14:04:56.157484 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="sg-core" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157495 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="sg-core" Nov 24 14:04:56 crc kubenswrapper[4970]: E1124 14:04:56.157516 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-central-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157525 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-central-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157787 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="proxy-httpd" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157812 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="sg-core" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157825 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-notification-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.157840 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" containerName="ceilometer-central-agent" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.162024 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.165720 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.165726 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.167455 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.178550 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.321913 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322028 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322069 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322092 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322234 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322275 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsrzx\" (UniqueName: \"kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322324 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.322414 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.424143 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.424719 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsrzx\" (UniqueName: \"kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.424781 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.424913 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.425019 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.425056 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.425098 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.425163 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.426133 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.426349 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.430285 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.430626 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.430989 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.432086 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.432265 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.449711 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsrzx\" (UniqueName: \"kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx\") pod \"ceilometer-0\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.485101 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:04:56 crc kubenswrapper[4970]: I1124 14:04:56.793116 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.031529 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.044726 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.046346 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.480821 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b90d420e-5ea1-46f4-8b51-ba8b1957143b" path="/var/lib/kubelet/pods/b90d420e-5ea1-46f4-8b51-ba8b1957143b/volumes" Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.778586 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerStarted","Data":"44ede77877357ee08ecc6e6f168350958779ffa65108089215b44c5d0a120931"} Nov 24 14:04:57 crc kubenswrapper[4970]: I1124 14:04:57.779069 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerStarted","Data":"f34b911d449e3ea399c1d43324f3722bd048321f1076235cba9237d5360a3d17"} Nov 24 14:04:58 crc kubenswrapper[4970]: I1124 14:04:58.128820 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:04:58 crc kubenswrapper[4970]: I1124 14:04:58.129201 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.198:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 14:04:58 crc kubenswrapper[4970]: I1124 14:04:58.788680 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerStarted","Data":"9175f91c3a656c967b02e953b3367f19daf3025fea90680a943e5cac9d5f30c7"} Nov 24 14:04:59 crc kubenswrapper[4970]: I1124 14:04:59.808940 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerStarted","Data":"a2199e6a9476642489181b3d4b2eb4ce05a63a5ced7271adef5a71049752a165"} Nov 24 14:05:01 crc kubenswrapper[4970]: I1124 14:05:01.838205 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerStarted","Data":"7962c2719faf0de17364f9edf5456f8b31c6b3ea525b101663a738a561c186f9"} Nov 24 14:05:01 crc kubenswrapper[4970]: I1124 14:05:01.838738 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:05:01 crc kubenswrapper[4970]: I1124 14:05:01.871386 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.801809924 podStartE2EDuration="5.871365456s" podCreationTimestamp="2025-11-24 14:04:56 +0000 UTC" firstStartedPulling="2025-11-24 14:04:57.046894583 +0000 UTC m=+1112.334651876" lastFinishedPulling="2025-11-24 14:05:01.116450085 +0000 UTC m=+1116.404207408" observedRunningTime="2025-11-24 14:05:01.862024148 +0000 UTC m=+1117.149781441" watchObservedRunningTime="2025-11-24 14:05:01.871365456 +0000 UTC m=+1117.159122749" Nov 24 14:05:03 crc kubenswrapper[4970]: E1124 14:05:03.671277 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc96d3bcf_7d2f_4946_a998_2567e91b275c.slice/crio-conmon-11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc96d3bcf_7d2f_4946_a998_2567e91b275c.slice/crio-11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90d420e_5ea1_46f4_8b51_ba8b1957143b.slice/crio-conmon-eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice/crio-06820e82571649efd821f88ca43bde4507e5f7eec90d3c2509d9951c865463cd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93ab7a4e_1f4e_49f3_8cbc_343e2c8aef66.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb87ce82_23c4_4610_a607_c648b7c88b81.slice/crio-conmon-c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb87ce82_23c4_4610_a607_c648b7c88b81.slice/crio-947b7a79252a07fc3896e23d26f0ef4cce595f8dc977c469fe7b37f714bbf543\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90d420e_5ea1_46f4_8b51_ba8b1957143b.slice/crio-8ec3f2aede6d8d078f7a6044d11f905a24cdc211b27852a69524243ad1cd6fad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc96d3bcf_7d2f_4946_a998_2567e91b275c.slice/crio-c84705250b890dc30563dcc3805ac60fecb1269e12856a206fa4efbed104d75c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb87ce82_23c4_4610_a607_c648b7c88b81.slice/crio-c6fb6c7c8932df0169518e4746f7e187c96fba0a3ff61efbe7c2369f7965999f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb87ce82_23c4_4610_a607_c648b7c88b81.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90d420e_5ea1_46f4_8b51_ba8b1957143b.slice/crio-eb2a06d60a04cad098bd2c573b7c84ba79f3da12fad5e0e23ece3b9c17a40ad3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb90d420e_5ea1_46f4_8b51_ba8b1957143b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod478cab1a_9f02_4f21_bae0_5f11ddc62665.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc96d3bcf_7d2f_4946_a998_2567e91b275c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ec89865_98e3_46d3_84cc_b4e5120b992f.slice/crio-fb9c85057444f17756ff32ec01ff1bae6723c326f9992bef7ff71bbe7e592b16\": RecentStats: unable to find data in memory cache]" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.856559 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.860137 4970 generic.go:334] "Generic (PLEG): container finished" podID="26a9ee71-436d-4a52-9685-ab16d64796d1" containerID="26e36d6d8d834f11c4fc7136e5c9bb504ac9e0413b4ead7d36b8eb75e0a031fd" exitCode=137 Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.860228 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"26a9ee71-436d-4a52-9685-ab16d64796d1","Type":"ContainerDied","Data":"26e36d6d8d834f11c4fc7136e5c9bb504ac9e0413b4ead7d36b8eb75e0a031fd"} Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.860266 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"26a9ee71-436d-4a52-9685-ab16d64796d1","Type":"ContainerDied","Data":"0c4364bb6a0ebfd6e25cbfa233bf9781afe929e32b8ac96443c5f0e92fc55e3d"} Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.860290 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c4364bb6a0ebfd6e25cbfa233bf9781afe929e32b8ac96443c5f0e92fc55e3d" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.862013 4970 generic.go:334] "Generic (PLEG): container finished" podID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerID="74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b" exitCode=137 Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.862048 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerDied","Data":"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b"} Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.862127 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"41bf37c2-7bb7-4d61-b856-b9550a89598d","Type":"ContainerDied","Data":"624fd0a3f3c4f0db5ba1a1199ee412463552a6d8a7102d7ce431d9ec4bb4a853"} Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.862147 4970 scope.go:117] "RemoveContainer" containerID="74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.862189 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.873444 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.890611 4970 scope.go:117] "RemoveContainer" containerID="64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.917688 4970 scope.go:117] "RemoveContainer" containerID="74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b" Nov 24 14:05:03 crc kubenswrapper[4970]: E1124 14:05:03.918172 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b\": container with ID starting with 74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b not found: ID does not exist" containerID="74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.918205 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b"} err="failed to get container status \"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b\": rpc error: code = NotFound desc = could not find container \"74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b\": container with ID starting with 74f3ed2385dd579c52a30221eac0e8fc5e86a69d1dc7a6c0d9ae772c71cfe47b not found: ID does not exist" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.918226 4970 scope.go:117] "RemoveContainer" containerID="64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef" Nov 24 14:05:03 crc kubenswrapper[4970]: E1124 14:05:03.918809 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef\": container with ID starting with 64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef not found: ID does not exist" containerID="64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef" Nov 24 14:05:03 crc kubenswrapper[4970]: I1124 14:05:03.918849 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef"} err="failed to get container status \"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef\": rpc error: code = NotFound desc = could not find container \"64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef\": container with ID starting with 64796984c13b6ae80d50fe65e7e5670017902f8dc7a31211f89da3155e3e15ef not found: ID does not exist" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.027807 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp9qg\" (UniqueName: \"kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg\") pod \"26a9ee71-436d-4a52-9685-ab16d64796d1\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.027921 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs\") pod \"41bf37c2-7bb7-4d61-b856-b9550a89598d\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028018 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle\") pod \"41bf37c2-7bb7-4d61-b856-b9550a89598d\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028045 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data\") pod \"41bf37c2-7bb7-4d61-b856-b9550a89598d\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028078 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd5gh\" (UniqueName: \"kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh\") pod \"41bf37c2-7bb7-4d61-b856-b9550a89598d\" (UID: \"41bf37c2-7bb7-4d61-b856-b9550a89598d\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028110 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle\") pod \"26a9ee71-436d-4a52-9685-ab16d64796d1\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028129 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data\") pod \"26a9ee71-436d-4a52-9685-ab16d64796d1\" (UID: \"26a9ee71-436d-4a52-9685-ab16d64796d1\") " Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028337 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs" (OuterVolumeSpecName: "logs") pod "41bf37c2-7bb7-4d61-b856-b9550a89598d" (UID: "41bf37c2-7bb7-4d61-b856-b9550a89598d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.028630 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41bf37c2-7bb7-4d61-b856-b9550a89598d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.034408 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg" (OuterVolumeSpecName: "kube-api-access-wp9qg") pod "26a9ee71-436d-4a52-9685-ab16d64796d1" (UID: "26a9ee71-436d-4a52-9685-ab16d64796d1"). InnerVolumeSpecName "kube-api-access-wp9qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.054295 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh" (OuterVolumeSpecName: "kube-api-access-sd5gh") pod "41bf37c2-7bb7-4d61-b856-b9550a89598d" (UID: "41bf37c2-7bb7-4d61-b856-b9550a89598d"). InnerVolumeSpecName "kube-api-access-sd5gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.056367 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26a9ee71-436d-4a52-9685-ab16d64796d1" (UID: "26a9ee71-436d-4a52-9685-ab16d64796d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.063471 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data" (OuterVolumeSpecName: "config-data") pod "26a9ee71-436d-4a52-9685-ab16d64796d1" (UID: "26a9ee71-436d-4a52-9685-ab16d64796d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.074363 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41bf37c2-7bb7-4d61-b856-b9550a89598d" (UID: "41bf37c2-7bb7-4d61-b856-b9550a89598d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.088806 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data" (OuterVolumeSpecName: "config-data") pod "41bf37c2-7bb7-4d61-b856-b9550a89598d" (UID: "41bf37c2-7bb7-4d61-b856-b9550a89598d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130194 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130233 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd5gh\" (UniqueName: \"kubernetes.io/projected/41bf37c2-7bb7-4d61-b856-b9550a89598d-kube-api-access-sd5gh\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130248 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130259 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a9ee71-436d-4a52-9685-ab16d64796d1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130278 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp9qg\" (UniqueName: \"kubernetes.io/projected/26a9ee71-436d-4a52-9685-ab16d64796d1-kube-api-access-wp9qg\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.130289 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41bf37c2-7bb7-4d61-b856-b9550a89598d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.207695 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.222507 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231015 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: E1124 14:05:04.231549 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-log" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231569 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-log" Nov 24 14:05:04 crc kubenswrapper[4970]: E1124 14:05:04.231613 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-metadata" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231623 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-metadata" Nov 24 14:05:04 crc kubenswrapper[4970]: E1124 14:05:04.231657 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a9ee71-436d-4a52-9685-ab16d64796d1" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231667 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a9ee71-436d-4a52-9685-ab16d64796d1" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231897 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a9ee71-436d-4a52-9685-ab16d64796d1" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231928 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-metadata" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.231943 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" containerName="nova-metadata-log" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.233236 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.236980 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.237288 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.238645 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.333758 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.333808 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.333954 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.334031 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2tbj\" (UniqueName: \"kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.334171 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.435278 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.435362 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.435385 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.435428 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.435457 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2tbj\" (UniqueName: \"kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.437057 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.440132 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.440703 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.440886 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.459484 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2tbj\" (UniqueName: \"kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj\") pod \"nova-metadata-0\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.558209 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.875105 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.925663 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.934118 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.956932 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.958360 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.960880 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.961009 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.972524 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:05:04 crc kubenswrapper[4970]: I1124 14:05:04.975183 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 14:05:05 crc kubenswrapper[4970]: W1124 14:05:05.051133 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeaae95d4_ac49_48bf_8a37_ba8aca50eff9.slice/crio-bed9b99e32331e43a16be5a6bdfdb0cb765a39bc5ceda1d7f0b4346557e55eb5 WatchSource:0}: Error finding container bed9b99e32331e43a16be5a6bdfdb0cb765a39bc5ceda1d7f0b4346557e55eb5: Status 404 returned error can't find the container with id bed9b99e32331e43a16be5a6bdfdb0cb765a39bc5ceda1d7f0b4346557e55eb5 Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.052049 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.058246 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.058277 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.058311 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8bkx\" (UniqueName: \"kubernetes.io/projected/5271e19c-5bd9-4c25-a6b9-a919f1445cca-kube-api-access-m8bkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.058347 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.058396 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.160062 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.160107 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.160150 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8bkx\" (UniqueName: \"kubernetes.io/projected/5271e19c-5bd9-4c25-a6b9-a919f1445cca-kube-api-access-m8bkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.160196 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.160269 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.165698 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.166201 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.166397 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.166741 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5271e19c-5bd9-4c25-a6b9-a919f1445cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.178253 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8bkx\" (UniqueName: \"kubernetes.io/projected/5271e19c-5bd9-4c25-a6b9-a919f1445cca-kube-api-access-m8bkx\") pod \"nova-cell1-novncproxy-0\" (UID: \"5271e19c-5bd9-4c25-a6b9-a919f1445cca\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.290250 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.489648 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a9ee71-436d-4a52-9685-ab16d64796d1" path="/var/lib/kubelet/pods/26a9ee71-436d-4a52-9685-ab16d64796d1/volumes" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.490307 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41bf37c2-7bb7-4d61-b856-b9550a89598d" path="/var/lib/kubelet/pods/41bf37c2-7bb7-4d61-b856-b9550a89598d/volumes" Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.532766 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.915734 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerStarted","Data":"e27324ee691ef282883c13aed7c96faa2c6e7112643f7c9371cc1231132a9207"} Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.915987 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerStarted","Data":"6297b317ce9b6fde1eeae8452b377bb42c0325bea6c8d2c94ce655fed7bfbaf5"} Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.915998 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerStarted","Data":"bed9b99e32331e43a16be5a6bdfdb0cb765a39bc5ceda1d7f0b4346557e55eb5"} Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.918559 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5271e19c-5bd9-4c25-a6b9-a919f1445cca","Type":"ContainerStarted","Data":"b51bf257fd528bbc29b70351f0609bc3f5392b10f49582b47551430b6b799d27"} Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.918611 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5271e19c-5bd9-4c25-a6b9-a919f1445cca","Type":"ContainerStarted","Data":"fd3bce3580ea1e9640002781e998628e36a30df88c1d2b9901b36958f3b6a742"} Nov 24 14:05:05 crc kubenswrapper[4970]: I1124 14:05:05.991453 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.9914337579999999 podStartE2EDuration="1.991433758s" podCreationTimestamp="2025-11-24 14:05:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:05.985776397 +0000 UTC m=+1121.273533690" watchObservedRunningTime="2025-11-24 14:05:05.991433758 +0000 UTC m=+1121.279191051" Nov 24 14:05:06 crc kubenswrapper[4970]: I1124 14:05:06.012956 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.012938417 podStartE2EDuration="2.012938417s" podCreationTimestamp="2025-11-24 14:05:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:06.01266254 +0000 UTC m=+1121.300419833" watchObservedRunningTime="2025-11-24 14:05:06.012938417 +0000 UTC m=+1121.300695710" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.043734 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.045235 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.048096 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.049849 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.951315 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:05:07 crc kubenswrapper[4970]: I1124 14:05:07.957137 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.105093 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.107751 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.126588 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.224453 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lgfx\" (UniqueName: \"kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.224570 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.224743 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.224852 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.224951 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.225060 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327046 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lgfx\" (UniqueName: \"kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327317 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327357 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327390 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327423 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.327460 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.328351 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.328380 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.328616 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.328923 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.329005 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.357311 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lgfx\" (UniqueName: \"kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx\") pod \"dnsmasq-dns-55bfb77665-7g8g8\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.432963 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.933020 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:05:08 crc kubenswrapper[4970]: W1124 14:05:08.933231 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69bf2307_7427_46de_89a0_18c31d0b0c73.slice/crio-60b466328a69432792adfb251ed837abf7869cb1490fbfaa2aa8245c55925268 WatchSource:0}: Error finding container 60b466328a69432792adfb251ed837abf7869cb1490fbfaa2aa8245c55925268: Status 404 returned error can't find the container with id 60b466328a69432792adfb251ed837abf7869cb1490fbfaa2aa8245c55925268 Nov 24 14:05:08 crc kubenswrapper[4970]: I1124 14:05:08.968745 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" event={"ID":"69bf2307-7427-46de-89a0-18c31d0b0c73","Type":"ContainerStarted","Data":"60b466328a69432792adfb251ed837abf7869cb1490fbfaa2aa8245c55925268"} Nov 24 14:05:09 crc kubenswrapper[4970]: I1124 14:05:09.558733 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:05:09 crc kubenswrapper[4970]: I1124 14:05:09.559066 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:05:09 crc kubenswrapper[4970]: I1124 14:05:09.976138 4970 generic.go:334] "Generic (PLEG): container finished" podID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerID="449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7" exitCode=0 Nov 24 14:05:09 crc kubenswrapper[4970]: I1124 14:05:09.976261 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" event={"ID":"69bf2307-7427-46de-89a0-18c31d0b0c73","Type":"ContainerDied","Data":"449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.085985 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.093540 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-central-agent" containerID="cri-o://44ede77877357ee08ecc6e6f168350958779ffa65108089215b44c5d0a120931" gracePeriod=30 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.094161 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="proxy-httpd" containerID="cri-o://7962c2719faf0de17364f9edf5456f8b31c6b3ea525b101663a738a561c186f9" gracePeriod=30 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.094223 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="sg-core" containerID="cri-o://a2199e6a9476642489181b3d4b2eb4ce05a63a5ced7271adef5a71049752a165" gracePeriod=30 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.094266 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-notification-agent" containerID="cri-o://9175f91c3a656c967b02e953b3367f19daf3025fea90680a943e5cac9d5f30c7" gracePeriod=30 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.291389 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.648678 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.989601 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" event={"ID":"69bf2307-7427-46de-89a0-18c31d0b0c73","Type":"ContainerStarted","Data":"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.990977 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995436 4970 generic.go:334] "Generic (PLEG): container finished" podID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerID="7962c2719faf0de17364f9edf5456f8b31c6b3ea525b101663a738a561c186f9" exitCode=0 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995464 4970 generic.go:334] "Generic (PLEG): container finished" podID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerID="a2199e6a9476642489181b3d4b2eb4ce05a63a5ced7271adef5a71049752a165" exitCode=2 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995478 4970 generic.go:334] "Generic (PLEG): container finished" podID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerID="9175f91c3a656c967b02e953b3367f19daf3025fea90680a943e5cac9d5f30c7" exitCode=0 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995490 4970 generic.go:334] "Generic (PLEG): container finished" podID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerID="44ede77877357ee08ecc6e6f168350958779ffa65108089215b44c5d0a120931" exitCode=0 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995805 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerDied","Data":"7962c2719faf0de17364f9edf5456f8b31c6b3ea525b101663a738a561c186f9"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995862 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerDied","Data":"a2199e6a9476642489181b3d4b2eb4ce05a63a5ced7271adef5a71049752a165"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995873 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerDied","Data":"9175f91c3a656c967b02e953b3367f19daf3025fea90680a943e5cac9d5f30c7"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995881 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerDied","Data":"44ede77877357ee08ecc6e6f168350958779ffa65108089215b44c5d0a120931"} Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995897 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-log" containerID="cri-o://3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9" gracePeriod=30 Nov 24 14:05:10 crc kubenswrapper[4970]: I1124 14:05:10.995958 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-api" containerID="cri-o://74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c" gracePeriod=30 Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.020302 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" podStartSLOduration=3.020281287 podStartE2EDuration="3.020281287s" podCreationTimestamp="2025-11-24 14:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:11.013328693 +0000 UTC m=+1126.301085986" watchObservedRunningTime="2025-11-24 14:05:11.020281287 +0000 UTC m=+1126.308038580" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.092708 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.181917 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182016 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182077 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsrzx\" (UniqueName: \"kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182134 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182203 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182231 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182281 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182369 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data\") pod \"9072b5e1-5490-431c-98a8-bbab5dff032a\" (UID: \"9072b5e1-5490-431c-98a8-bbab5dff032a\") " Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182464 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182869 4970 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.182916 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.188020 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts" (OuterVolumeSpecName: "scripts") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.188690 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx" (OuterVolumeSpecName: "kube-api-access-tsrzx") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "kube-api-access-tsrzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.204880 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.204936 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.204984 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.205783 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.205869 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e" gracePeriod=600 Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.248734 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.252105 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.284152 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.284184 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsrzx\" (UniqueName: \"kubernetes.io/projected/9072b5e1-5490-431c-98a8-bbab5dff032a-kube-api-access-tsrzx\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.284194 4970 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.284202 4970 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9072b5e1-5490-431c-98a8-bbab5dff032a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.284210 4970 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.288714 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.311201 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data" (OuterVolumeSpecName: "config-data") pod "9072b5e1-5490-431c-98a8-bbab5dff032a" (UID: "9072b5e1-5490-431c-98a8-bbab5dff032a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.386171 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:11 crc kubenswrapper[4970]: I1124 14:05:11.386207 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9072b5e1-5490-431c-98a8-bbab5dff032a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.006318 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e" exitCode=0 Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.006370 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e"} Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.007870 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57"} Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.007915 4970 scope.go:117] "RemoveContainer" containerID="97e436849708758ef637457aa47282c99d14376c5863acf143ee0e8b9ea7fd4c" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.010176 4970 generic.go:334] "Generic (PLEG): container finished" podID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerID="3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9" exitCode=143 Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.010287 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerDied","Data":"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9"} Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.013796 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9072b5e1-5490-431c-98a8-bbab5dff032a","Type":"ContainerDied","Data":"f34b911d449e3ea399c1d43324f3722bd048321f1076235cba9237d5360a3d17"} Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.013865 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.058164 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.063519 4970 scope.go:117] "RemoveContainer" containerID="7962c2719faf0de17364f9edf5456f8b31c6b3ea525b101663a738a561c186f9" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.070640 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.085033 4970 scope.go:117] "RemoveContainer" containerID="a2199e6a9476642489181b3d4b2eb4ce05a63a5ced7271adef5a71049752a165" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.091784 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:12 crc kubenswrapper[4970]: E1124 14:05:12.092176 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-notification-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.092192 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-notification-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: E1124 14:05:12.092221 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="sg-core" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.092227 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="sg-core" Nov 24 14:05:12 crc kubenswrapper[4970]: E1124 14:05:12.092240 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="proxy-httpd" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.092248 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="proxy-httpd" Nov 24 14:05:12 crc kubenswrapper[4970]: E1124 14:05:12.092260 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-central-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.092266 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-central-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.094334 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-notification-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.094364 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="ceilometer-central-agent" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.094376 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="proxy-httpd" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.094393 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" containerName="sg-core" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.096019 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.099280 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.099305 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.099421 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.110874 4970 scope.go:117] "RemoveContainer" containerID="9175f91c3a656c967b02e953b3367f19daf3025fea90680a943e5cac9d5f30c7" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.112709 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.154961 4970 scope.go:117] "RemoveContainer" containerID="44ede77877357ee08ecc6e6f168350958779ffa65108089215b44c5d0a120931" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200347 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks8v8\" (UniqueName: \"kubernetes.io/projected/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-kube-api-access-ks8v8\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200393 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-run-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200427 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200478 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-config-data\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200516 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200534 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200568 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-scripts\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.200641 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-log-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302192 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-run-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302264 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302372 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-config-data\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302459 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302490 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302534 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-scripts\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302556 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-log-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.302661 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks8v8\" (UniqueName: \"kubernetes.io/projected/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-kube-api-access-ks8v8\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.303074 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-run-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.303668 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-log-httpd\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.311066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.313271 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-scripts\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.313315 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.313664 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.316224 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-config-data\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.323414 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks8v8\" (UniqueName: \"kubernetes.io/projected/de278d27-bc30-4d46-a4b3-3019fb4e6e3c-kube-api-access-ks8v8\") pod \"ceilometer-0\" (UID: \"de278d27-bc30-4d46-a4b3-3019fb4e6e3c\") " pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.423688 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 14:05:12 crc kubenswrapper[4970]: I1124 14:05:12.887910 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 14:05:13 crc kubenswrapper[4970]: I1124 14:05:13.026648 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de278d27-bc30-4d46-a4b3-3019fb4e6e3c","Type":"ContainerStarted","Data":"69de63650682e0080a0b86133893f89223e9fb03eb5cb699de19f1588d0ad5cc"} Nov 24 14:05:13 crc kubenswrapper[4970]: I1124 14:05:13.482804 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9072b5e1-5490-431c-98a8-bbab5dff032a" path="/var/lib/kubelet/pods/9072b5e1-5490-431c-98a8-bbab5dff032a/volumes" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.043230 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de278d27-bc30-4d46-a4b3-3019fb4e6e3c","Type":"ContainerStarted","Data":"21d85396b2553ad6050b2711a84ccdd92e54e5c4c26ec5ce7eafdcfeca704d5c"} Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.561090 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.562875 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.678865 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.754252 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs\") pod \"c78e24d9-cd2a-4832-a066-c9692048ad3d\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.754355 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp8vz\" (UniqueName: \"kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz\") pod \"c78e24d9-cd2a-4832-a066-c9692048ad3d\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.754440 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle\") pod \"c78e24d9-cd2a-4832-a066-c9692048ad3d\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.754483 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data\") pod \"c78e24d9-cd2a-4832-a066-c9692048ad3d\" (UID: \"c78e24d9-cd2a-4832-a066-c9692048ad3d\") " Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.754897 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs" (OuterVolumeSpecName: "logs") pod "c78e24d9-cd2a-4832-a066-c9692048ad3d" (UID: "c78e24d9-cd2a-4832-a066-c9692048ad3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.762466 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz" (OuterVolumeSpecName: "kube-api-access-wp8vz") pod "c78e24d9-cd2a-4832-a066-c9692048ad3d" (UID: "c78e24d9-cd2a-4832-a066-c9692048ad3d"). InnerVolumeSpecName "kube-api-access-wp8vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.783163 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c78e24d9-cd2a-4832-a066-c9692048ad3d" (UID: "c78e24d9-cd2a-4832-a066-c9692048ad3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.819653 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data" (OuterVolumeSpecName: "config-data") pod "c78e24d9-cd2a-4832-a066-c9692048ad3d" (UID: "c78e24d9-cd2a-4832-a066-c9692048ad3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.856265 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.856289 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c78e24d9-cd2a-4832-a066-c9692048ad3d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.856297 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c78e24d9-cd2a-4832-a066-c9692048ad3d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:14 crc kubenswrapper[4970]: I1124 14:05:14.856306 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp8vz\" (UniqueName: \"kubernetes.io/projected/c78e24d9-cd2a-4832-a066-c9692048ad3d-kube-api-access-wp8vz\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.060755 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de278d27-bc30-4d46-a4b3-3019fb4e6e3c","Type":"ContainerStarted","Data":"085492c426cd9941e71aaeb42770f683d2405fddcaafc319cfa5348e77828d71"} Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.064746 4970 generic.go:334] "Generic (PLEG): container finished" podID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerID="74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c" exitCode=0 Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.065762 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.066183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerDied","Data":"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c"} Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.066229 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c78e24d9-cd2a-4832-a066-c9692048ad3d","Type":"ContainerDied","Data":"c8961bdf6451482004b799bb9056c29f4ebb66cf800ea765724cf1e9413bc426"} Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.066251 4970 scope.go:117] "RemoveContainer" containerID="74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.089345 4970 scope.go:117] "RemoveContainer" containerID="3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.109328 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.122072 4970 scope.go:117] "RemoveContainer" containerID="74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c" Nov 24 14:05:15 crc kubenswrapper[4970]: E1124 14:05:15.122547 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c\": container with ID starting with 74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c not found: ID does not exist" containerID="74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.122605 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c"} err="failed to get container status \"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c\": rpc error: code = NotFound desc = could not find container \"74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c\": container with ID starting with 74452550e07105f45f5647bf5a12a6e04779a35a8801743eda5322b0a745786c not found: ID does not exist" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.122638 4970 scope.go:117] "RemoveContainer" containerID="3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9" Nov 24 14:05:15 crc kubenswrapper[4970]: E1124 14:05:15.122860 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9\": container with ID starting with 3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9 not found: ID does not exist" containerID="3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.122889 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9"} err="failed to get container status \"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9\": rpc error: code = NotFound desc = could not find container \"3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9\": container with ID starting with 3d39ea2ca46b366ae1ab1fe416ae591f96abe7f414d77215abf2fe0270c6e5c9 not found: ID does not exist" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.126914 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.139621 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:15 crc kubenswrapper[4970]: E1124 14:05:15.140143 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-log" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.140163 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-log" Nov 24 14:05:15 crc kubenswrapper[4970]: E1124 14:05:15.140198 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-api" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.140206 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-api" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.140450 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-log" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.140482 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" containerName="nova-api-api" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.141628 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.144126 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.144507 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.144692 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.171395 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268031 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcbzm\" (UniqueName: \"kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268122 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268158 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268307 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268385 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.268410 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.291086 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.318922 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369646 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369689 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369798 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369834 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369848 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.369889 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcbzm\" (UniqueName: \"kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.370407 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.374101 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.374430 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.377142 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.378621 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.391363 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcbzm\" (UniqueName: \"kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm\") pod \"nova-api-0\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.464742 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.489931 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c78e24d9-cd2a-4832-a066-c9692048ad3d" path="/var/lib/kubelet/pods/c78e24d9-cd2a-4832-a066-c9692048ad3d/volumes" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.577756 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.578046 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:15 crc kubenswrapper[4970]: W1124 14:05:15.958276 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ecca8b8_1c75_4f1e_b914_f1197bc45180.slice/crio-d0d2fd46bee2134ceab619e684d3d920d233b1f331e03526531dff89c6146ee2 WatchSource:0}: Error finding container d0d2fd46bee2134ceab619e684d3d920d233b1f331e03526531dff89c6146ee2: Status 404 returned error can't find the container with id d0d2fd46bee2134ceab619e684d3d920d233b1f331e03526531dff89c6146ee2 Nov 24 14:05:15 crc kubenswrapper[4970]: I1124 14:05:15.966252 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.082337 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerStarted","Data":"d0d2fd46bee2134ceab619e684d3d920d233b1f331e03526531dff89c6146ee2"} Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.089370 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de278d27-bc30-4d46-a4b3-3019fb4e6e3c","Type":"ContainerStarted","Data":"9c8eb6db8509b26dabf7ccf769e8f3d126e5bcd68c7e56209017b525d0dfb15d"} Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.105296 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.404170 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-chg79"] Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.409142 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.411131 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.411630 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.429021 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-chg79"] Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.491772 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.491901 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.493595 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.494629 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkrwn\" (UniqueName: \"kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.597014 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkrwn\" (UniqueName: \"kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.597094 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.597149 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.597169 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.603818 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.605187 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.607881 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.615306 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkrwn\" (UniqueName: \"kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn\") pod \"nova-cell1-cell-mapping-chg79\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:16 crc kubenswrapper[4970]: I1124 14:05:16.731242 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:17 crc kubenswrapper[4970]: I1124 14:05:17.107361 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerStarted","Data":"901ba548954f6d329c0e897de5645751f48c83f32674e374ff40751f5382a603"} Nov 24 14:05:17 crc kubenswrapper[4970]: I1124 14:05:17.107756 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerStarted","Data":"25c8c83aa7fce19e142eb01a3da6bc7ddd12d2ce8cd3839e478670b0a9266e5a"} Nov 24 14:05:17 crc kubenswrapper[4970]: I1124 14:05:17.142167 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.14214363 podStartE2EDuration="2.14214363s" podCreationTimestamp="2025-11-24 14:05:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:17.129967358 +0000 UTC m=+1132.417724651" watchObservedRunningTime="2025-11-24 14:05:17.14214363 +0000 UTC m=+1132.429900943" Nov 24 14:05:17 crc kubenswrapper[4970]: W1124 14:05:17.227450 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode87b7aad_6180_461c_981c_f1a1c8be8beb.slice/crio-91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea WatchSource:0}: Error finding container 91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea: Status 404 returned error can't find the container with id 91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea Nov 24 14:05:17 crc kubenswrapper[4970]: I1124 14:05:17.235940 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-chg79"] Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.127367 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-chg79" event={"ID":"e87b7aad-6180-461c-981c-f1a1c8be8beb","Type":"ContainerStarted","Data":"853d1adce379e23a214bf4f8822859a9feae49e1dca8e86f02431eb9ee9b0c2e"} Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.127826 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-chg79" event={"ID":"e87b7aad-6180-461c-981c-f1a1c8be8beb","Type":"ContainerStarted","Data":"91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea"} Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.133959 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"de278d27-bc30-4d46-a4b3-3019fb4e6e3c","Type":"ContainerStarted","Data":"eaa95ee0126dafcbfadfbfd0d84cc17355b21de1d2e5bad1bc772ebe69743e88"} Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.180532 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-chg79" podStartSLOduration=2.180455814 podStartE2EDuration="2.180455814s" podCreationTimestamp="2025-11-24 14:05:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:18.166561165 +0000 UTC m=+1133.454318498" watchObservedRunningTime="2025-11-24 14:05:18.180455814 +0000 UTC m=+1133.468213107" Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.202797 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.081105071 podStartE2EDuration="6.202775375s" podCreationTimestamp="2025-11-24 14:05:12 +0000 UTC" firstStartedPulling="2025-11-24 14:05:12.890599695 +0000 UTC m=+1128.178356988" lastFinishedPulling="2025-11-24 14:05:17.012269999 +0000 UTC m=+1132.300027292" observedRunningTime="2025-11-24 14:05:18.195513853 +0000 UTC m=+1133.483271166" watchObservedRunningTime="2025-11-24 14:05:18.202775375 +0000 UTC m=+1133.490532678" Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.434861 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.511723 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:05:18 crc kubenswrapper[4970]: I1124 14:05:18.512024 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="dnsmasq-dns" containerID="cri-o://a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584" gracePeriod=10 Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.121303 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.145548 4970 generic.go:334] "Generic (PLEG): container finished" podID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerID="a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584" exitCode=0 Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.146493 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" event={"ID":"d498d983-283c-4e6d-a2bb-0b143562a2c1","Type":"ContainerDied","Data":"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584"} Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.146518 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.146560 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" event={"ID":"d498d983-283c-4e6d-a2bb-0b143562a2c1","Type":"ContainerDied","Data":"c63bb66c05e0ff3d079f9997282b4e9153c74b093f27389a3427adf1455ce4fd"} Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.146617 4970 scope.go:117] "RemoveContainer" containerID="a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.146813 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.176188 4970 scope.go:117] "RemoveContainer" containerID="485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.236735 4970 scope.go:117] "RemoveContainer" containerID="a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584" Nov 24 14:05:19 crc kubenswrapper[4970]: E1124 14:05:19.240416 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584\": container with ID starting with a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584 not found: ID does not exist" containerID="a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.240455 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584"} err="failed to get container status \"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584\": rpc error: code = NotFound desc = could not find container \"a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584\": container with ID starting with a9ba259d155d4e7730a8656b4d1fd421f68bbf0dcd45da32dc7470065a7f5584 not found: ID does not exist" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.240476 4970 scope.go:117] "RemoveContainer" containerID="485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e" Nov 24 14:05:19 crc kubenswrapper[4970]: E1124 14:05:19.241073 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e\": container with ID starting with 485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e not found: ID does not exist" containerID="485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.241117 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e"} err="failed to get container status \"485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e\": rpc error: code = NotFound desc = could not find container \"485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e\": container with ID starting with 485f13db6c36d0e49587bdd471a3b4cb595c3b329083170dd190329778cb353e not found: ID does not exist" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.310375 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.310471 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.310638 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b67nx\" (UniqueName: \"kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.310788 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.310943 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.311021 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb\") pod \"d498d983-283c-4e6d-a2bb-0b143562a2c1\" (UID: \"d498d983-283c-4e6d-a2bb-0b143562a2c1\") " Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.316422 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx" (OuterVolumeSpecName: "kube-api-access-b67nx") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "kube-api-access-b67nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.359601 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.369336 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.375095 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.380083 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config" (OuterVolumeSpecName: "config") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.381062 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d498d983-283c-4e6d-a2bb-0b143562a2c1" (UID: "d498d983-283c-4e6d-a2bb-0b143562a2c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414791 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b67nx\" (UniqueName: \"kubernetes.io/projected/d498d983-283c-4e6d-a2bb-0b143562a2c1-kube-api-access-b67nx\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414825 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414835 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414844 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414852 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.414861 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d498d983-283c-4e6d-a2bb-0b143562a2c1-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.486905 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:05:19 crc kubenswrapper[4970]: I1124 14:05:19.498103 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64dbf5859c-2k4gm"] Nov 24 14:05:21 crc kubenswrapper[4970]: I1124 14:05:21.483295 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" path="/var/lib/kubelet/pods/d498d983-283c-4e6d-a2bb-0b143562a2c1/volumes" Nov 24 14:05:23 crc kubenswrapper[4970]: I1124 14:05:23.192346 4970 generic.go:334] "Generic (PLEG): container finished" podID="e87b7aad-6180-461c-981c-f1a1c8be8beb" containerID="853d1adce379e23a214bf4f8822859a9feae49e1dca8e86f02431eb9ee9b0c2e" exitCode=0 Nov 24 14:05:23 crc kubenswrapper[4970]: I1124 14:05:23.192499 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-chg79" event={"ID":"e87b7aad-6180-461c-981c-f1a1c8be8beb","Type":"ContainerDied","Data":"853d1adce379e23a214bf4f8822859a9feae49e1dca8e86f02431eb9ee9b0c2e"} Nov 24 14:05:23 crc kubenswrapper[4970]: I1124 14:05:23.977241 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-64dbf5859c-2k4gm" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: i/o timeout" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.563392 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.565914 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.569164 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.621325 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.719546 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkrwn\" (UniqueName: \"kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn\") pod \"e87b7aad-6180-461c-981c-f1a1c8be8beb\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.719980 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data\") pod \"e87b7aad-6180-461c-981c-f1a1c8be8beb\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.720093 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts\") pod \"e87b7aad-6180-461c-981c-f1a1c8be8beb\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.720119 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle\") pod \"e87b7aad-6180-461c-981c-f1a1c8be8beb\" (UID: \"e87b7aad-6180-461c-981c-f1a1c8be8beb\") " Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.725793 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn" (OuterVolumeSpecName: "kube-api-access-hkrwn") pod "e87b7aad-6180-461c-981c-f1a1c8be8beb" (UID: "e87b7aad-6180-461c-981c-f1a1c8be8beb"). InnerVolumeSpecName "kube-api-access-hkrwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.725875 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts" (OuterVolumeSpecName: "scripts") pod "e87b7aad-6180-461c-981c-f1a1c8be8beb" (UID: "e87b7aad-6180-461c-981c-f1a1c8be8beb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.746812 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e87b7aad-6180-461c-981c-f1a1c8be8beb" (UID: "e87b7aad-6180-461c-981c-f1a1c8be8beb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.747542 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data" (OuterVolumeSpecName: "config-data") pod "e87b7aad-6180-461c-981c-f1a1c8be8beb" (UID: "e87b7aad-6180-461c-981c-f1a1c8be8beb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.821938 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.822202 4970 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.822264 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e87b7aad-6180-461c-981c-f1a1c8be8beb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:24 crc kubenswrapper[4970]: I1124 14:05:24.822333 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkrwn\" (UniqueName: \"kubernetes.io/projected/e87b7aad-6180-461c-981c-f1a1c8be8beb-kube-api-access-hkrwn\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.222285 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-chg79" Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.222283 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-chg79" event={"ID":"e87b7aad-6180-461c-981c-f1a1c8be8beb","Type":"ContainerDied","Data":"91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea"} Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.222846 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91e08800dfde976da7e7366e85faf62354d2ea0a4c05111f0ba10639375f53ea" Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.228822 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.406555 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.406887 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-log" containerID="cri-o://25c8c83aa7fce19e142eb01a3da6bc7ddd12d2ce8cd3839e478670b0a9266e5a" gracePeriod=30 Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.407419 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-api" containerID="cri-o://901ba548954f6d329c0e897de5645751f48c83f32674e374ff40751f5382a603" gracePeriod=30 Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.421722 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.421969 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerName="nova-scheduler-scheduler" containerID="cri-o://e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" gracePeriod=30 Nov 24 14:05:25 crc kubenswrapper[4970]: I1124 14:05:25.437456 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:26 crc kubenswrapper[4970]: E1124 14:05:26.029177 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:05:26 crc kubenswrapper[4970]: E1124 14:05:26.033327 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:05:26 crc kubenswrapper[4970]: E1124 14:05:26.035383 4970 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 14:05:26 crc kubenswrapper[4970]: E1124 14:05:26.035433 4970 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerName="nova-scheduler-scheduler" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.232539 4970 generic.go:334] "Generic (PLEG): container finished" podID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerID="901ba548954f6d329c0e897de5645751f48c83f32674e374ff40751f5382a603" exitCode=0 Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.232568 4970 generic.go:334] "Generic (PLEG): container finished" podID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerID="25c8c83aa7fce19e142eb01a3da6bc7ddd12d2ce8cd3839e478670b0a9266e5a" exitCode=143 Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.232627 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerDied","Data":"901ba548954f6d329c0e897de5645751f48c83f32674e374ff40751f5382a603"} Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.232662 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerDied","Data":"25c8c83aa7fce19e142eb01a3da6bc7ddd12d2ce8cd3839e478670b0a9266e5a"} Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.493999 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.653980 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.654311 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.654393 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.654509 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcbzm\" (UniqueName: \"kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.654543 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.654569 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data\") pod \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\" (UID: \"1ecca8b8-1c75-4f1e-b914-f1197bc45180\") " Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.655089 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs" (OuterVolumeSpecName: "logs") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.659510 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm" (OuterVolumeSpecName: "kube-api-access-hcbzm") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "kube-api-access-hcbzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.681283 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data" (OuterVolumeSpecName: "config-data") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.688598 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.703933 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.711571 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1ecca8b8-1c75-4f1e-b914-f1197bc45180" (UID: "1ecca8b8-1c75-4f1e-b914-f1197bc45180"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757064 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757104 4970 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757115 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ecca8b8-1c75-4f1e-b914-f1197bc45180-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757126 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcbzm\" (UniqueName: \"kubernetes.io/projected/1ecca8b8-1c75-4f1e-b914-f1197bc45180-kube-api-access-hcbzm\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757138 4970 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:26 crc kubenswrapper[4970]: I1124 14:05:26.757146 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ecca8b8-1c75-4f1e-b914-f1197bc45180-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.242247 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1ecca8b8-1c75-4f1e-b914-f1197bc45180","Type":"ContainerDied","Data":"d0d2fd46bee2134ceab619e684d3d920d233b1f331e03526531dff89c6146ee2"} Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.242313 4970 scope.go:117] "RemoveContainer" containerID="901ba548954f6d329c0e897de5645751f48c83f32674e374ff40751f5382a603" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.242263 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.242364 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" containerID="cri-o://6297b317ce9b6fde1eeae8452b377bb42c0325bea6c8d2c94ce655fed7bfbaf5" gracePeriod=30 Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.242668 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" containerID="cri-o://e27324ee691ef282883c13aed7c96faa2c6e7112643f7c9371cc1231132a9207" gracePeriod=30 Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.277870 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.285356 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.286494 4970 scope.go:117] "RemoveContainer" containerID="25c8c83aa7fce19e142eb01a3da6bc7ddd12d2ce8cd3839e478670b0a9266e5a" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.321505 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:27 crc kubenswrapper[4970]: E1124 14:05:27.321974 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="init" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.321999 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="init" Nov 24 14:05:27 crc kubenswrapper[4970]: E1124 14:05:27.322020 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="dnsmasq-dns" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322029 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="dnsmasq-dns" Nov 24 14:05:27 crc kubenswrapper[4970]: E1124 14:05:27.322040 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87b7aad-6180-461c-981c-f1a1c8be8beb" containerName="nova-manage" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322048 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87b7aad-6180-461c-981c-f1a1c8be8beb" containerName="nova-manage" Nov 24 14:05:27 crc kubenswrapper[4970]: E1124 14:05:27.322064 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-api" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322071 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-api" Nov 24 14:05:27 crc kubenswrapper[4970]: E1124 14:05:27.322093 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-log" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322100 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-log" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322323 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d498d983-283c-4e6d-a2bb-0b143562a2c1" containerName="dnsmasq-dns" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322347 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87b7aad-6180-461c-981c-f1a1c8be8beb" containerName="nova-manage" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322359 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-log" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.322373 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" containerName="nova-api-api" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.323663 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.327118 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.327250 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.327307 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.345233 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469472 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e7b6fd-92df-4c42-b84e-151948459084-logs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469542 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469701 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469738 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-public-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469810 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg9l4\" (UniqueName: \"kubernetes.io/projected/05e7b6fd-92df-4c42-b84e-151948459084-kube-api-access-sg9l4\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.469849 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-config-data\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.482267 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ecca8b8-1c75-4f1e-b914-f1197bc45180" path="/var/lib/kubelet/pods/1ecca8b8-1c75-4f1e-b914-f1197bc45180/volumes" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.571909 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.571982 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-public-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.572192 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg9l4\" (UniqueName: \"kubernetes.io/projected/05e7b6fd-92df-4c42-b84e-151948459084-kube-api-access-sg9l4\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.572224 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-config-data\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.572297 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e7b6fd-92df-4c42-b84e-151948459084-logs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.572321 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.574017 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05e7b6fd-92df-4c42-b84e-151948459084-logs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.578629 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-internal-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.581008 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-config-data\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.582629 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-public-tls-certs\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.589726 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e7b6fd-92df-4c42-b84e-151948459084-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.591909 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg9l4\" (UniqueName: \"kubernetes.io/projected/05e7b6fd-92df-4c42-b84e-151948459084-kube-api-access-sg9l4\") pod \"nova-api-0\" (UID: \"05e7b6fd-92df-4c42-b84e-151948459084\") " pod="openstack/nova-api-0" Nov 24 14:05:27 crc kubenswrapper[4970]: I1124 14:05:27.693548 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 14:05:28 crc kubenswrapper[4970]: I1124 14:05:28.135294 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 14:05:28 crc kubenswrapper[4970]: W1124 14:05:28.143550 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05e7b6fd_92df_4c42_b84e_151948459084.slice/crio-1f49170c266bd5d7d5374e41b1e98bef6e7cf2cfe079ca9278560f368323c97f WatchSource:0}: Error finding container 1f49170c266bd5d7d5374e41b1e98bef6e7cf2cfe079ca9278560f368323c97f: Status 404 returned error can't find the container with id 1f49170c266bd5d7d5374e41b1e98bef6e7cf2cfe079ca9278560f368323c97f Nov 24 14:05:28 crc kubenswrapper[4970]: I1124 14:05:28.258210 4970 generic.go:334] "Generic (PLEG): container finished" podID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerID="6297b317ce9b6fde1eeae8452b377bb42c0325bea6c8d2c94ce655fed7bfbaf5" exitCode=143 Nov 24 14:05:28 crc kubenswrapper[4970]: I1124 14:05:28.258285 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerDied","Data":"6297b317ce9b6fde1eeae8452b377bb42c0325bea6c8d2c94ce655fed7bfbaf5"} Nov 24 14:05:28 crc kubenswrapper[4970]: I1124 14:05:28.260102 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05e7b6fd-92df-4c42-b84e-151948459084","Type":"ContainerStarted","Data":"1f49170c266bd5d7d5374e41b1e98bef6e7cf2cfe079ca9278560f368323c97f"} Nov 24 14:05:29 crc kubenswrapper[4970]: I1124 14:05:29.484120 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05e7b6fd-92df-4c42-b84e-151948459084","Type":"ContainerStarted","Data":"2a8ebc56e1c80831297faa43782e63ccb10d68738d8f2e03ca23c97ebc50e7c3"} Nov 24 14:05:29 crc kubenswrapper[4970]: I1124 14:05:29.484546 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"05e7b6fd-92df-4c42-b84e-151948459084","Type":"ContainerStarted","Data":"a8908dbd5b2496faca5013d2ace3b26c4577941c2e666106c03c84519a915566"} Nov 24 14:05:29 crc kubenswrapper[4970]: I1124 14:05:29.503283 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.50326186 podStartE2EDuration="2.50326186s" podCreationTimestamp="2025-11-24 14:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:29.497328393 +0000 UTC m=+1144.785085686" watchObservedRunningTime="2025-11-24 14:05:29.50326186 +0000 UTC m=+1144.791019153" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.308456 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.396187 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:58946->10.217.0.200:8775: read: connection reset by peer" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.396192 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": read tcp 10.217.0.2:58962->10.217.0.200:8775: read: connection reset by peer" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.426446 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-656mn\" (UniqueName: \"kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn\") pod \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.426561 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data\") pod \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.426681 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle\") pod \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\" (UID: \"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.432409 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn" (OuterVolumeSpecName: "kube-api-access-656mn") pod "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" (UID: "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5"). InnerVolumeSpecName "kube-api-access-656mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.461724 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data" (OuterVolumeSpecName: "config-data") pod "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" (UID: "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.464837 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" (UID: "5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.488091 4970 generic.go:334] "Generic (PLEG): container finished" podID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerID="e27324ee691ef282883c13aed7c96faa2c6e7112643f7c9371cc1231132a9207" exitCode=0 Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.488156 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerDied","Data":"e27324ee691ef282883c13aed7c96faa2c6e7112643f7c9371cc1231132a9207"} Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.489985 4970 generic.go:334] "Generic (PLEG): container finished" podID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" exitCode=0 Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.490040 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.490089 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5","Type":"ContainerDied","Data":"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946"} Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.491955 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5","Type":"ContainerDied","Data":"06bcb6ed4193114135b9f95b5d183fb77c38b8d374698f898b60658b4c04e3cc"} Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.491985 4970 scope.go:117] "RemoveContainer" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.529338 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-656mn\" (UniqueName: \"kubernetes.io/projected/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-kube-api-access-656mn\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.529678 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.529693 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.550207 4970 scope.go:117] "RemoveContainer" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" Nov 24 14:05:30 crc kubenswrapper[4970]: E1124 14:05:30.550642 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946\": container with ID starting with e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946 not found: ID does not exist" containerID="e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.550674 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946"} err="failed to get container status \"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946\": rpc error: code = NotFound desc = could not find container \"e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946\": container with ID starting with e119c8a9d5f52bbd058ec7419cd261ffce14393ed8236c410a9433e0968b5946 not found: ID does not exist" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.555396 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.572143 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.585212 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:30 crc kubenswrapper[4970]: E1124 14:05:30.587428 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerName="nova-scheduler-scheduler" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.587445 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerName="nova-scheduler-scheduler" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.587669 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" containerName="nova-scheduler-scheduler" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.591795 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.591955 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.602021 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.735035 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.735125 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb8z8\" (UniqueName: \"kubernetes.io/projected/92b5c70e-7c5c-465f-a44c-adc0aed6c664-kube-api-access-gb8z8\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.735269 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-config-data\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.803225 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.837668 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-config-data\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.837816 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.837859 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb8z8\" (UniqueName: \"kubernetes.io/projected/92b5c70e-7c5c-465f-a44c-adc0aed6c664-kube-api-access-gb8z8\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.843458 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.845091 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92b5c70e-7c5c-465f-a44c-adc0aed6c664-config-data\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.855474 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb8z8\" (UniqueName: \"kubernetes.io/projected/92b5c70e-7c5c-465f-a44c-adc0aed6c664-kube-api-access-gb8z8\") pod \"nova-scheduler-0\" (UID: \"92b5c70e-7c5c-465f-a44c-adc0aed6c664\") " pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.917308 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.939190 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs\") pod \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.939236 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle\") pod \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.939271 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs\") pod \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.939408 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2tbj\" (UniqueName: \"kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj\") pod \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.939533 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data\") pod \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\" (UID: \"eaae95d4-ac49-48bf-8a37-ba8aca50eff9\") " Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.940662 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs" (OuterVolumeSpecName: "logs") pod "eaae95d4-ac49-48bf-8a37-ba8aca50eff9" (UID: "eaae95d4-ac49-48bf-8a37-ba8aca50eff9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.943387 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj" (OuterVolumeSpecName: "kube-api-access-q2tbj") pod "eaae95d4-ac49-48bf-8a37-ba8aca50eff9" (UID: "eaae95d4-ac49-48bf-8a37-ba8aca50eff9"). InnerVolumeSpecName "kube-api-access-q2tbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.971728 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaae95d4-ac49-48bf-8a37-ba8aca50eff9" (UID: "eaae95d4-ac49-48bf-8a37-ba8aca50eff9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:30 crc kubenswrapper[4970]: I1124 14:05:30.995352 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data" (OuterVolumeSpecName: "config-data") pod "eaae95d4-ac49-48bf-8a37-ba8aca50eff9" (UID: "eaae95d4-ac49-48bf-8a37-ba8aca50eff9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.031115 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "eaae95d4-ac49-48bf-8a37-ba8aca50eff9" (UID: "eaae95d4-ac49-48bf-8a37-ba8aca50eff9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.041329 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2tbj\" (UniqueName: \"kubernetes.io/projected/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-kube-api-access-q2tbj\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.041364 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.041413 4970 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.041422 4970 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.041430 4970 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaae95d4-ac49-48bf-8a37-ba8aca50eff9-logs\") on node \"crc\" DevicePath \"\"" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.377017 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.493222 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5" path="/var/lib/kubelet/pods/5b7d70bc-1cd7-47a6-93e9-7c3c64b559b5/volumes" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.524643 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eaae95d4-ac49-48bf-8a37-ba8aca50eff9","Type":"ContainerDied","Data":"bed9b99e32331e43a16be5a6bdfdb0cb765a39bc5ceda1d7f0b4346557e55eb5"} Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.524728 4970 scope.go:117] "RemoveContainer" containerID="e27324ee691ef282883c13aed7c96faa2c6e7112643f7c9371cc1231132a9207" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.525977 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.529396 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92b5c70e-7c5c-465f-a44c-adc0aed6c664","Type":"ContainerStarted","Data":"f668843da139a84c2a33ebd64979d8873d96e5b9175828bf94193ccaad0d2ad7"} Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.578670 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.581678 4970 scope.go:117] "RemoveContainer" containerID="6297b317ce9b6fde1eeae8452b377bb42c0325bea6c8d2c94ce655fed7bfbaf5" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.586177 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.597553 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:31 crc kubenswrapper[4970]: E1124 14:05:31.598012 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.598028 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" Nov 24 14:05:31 crc kubenswrapper[4970]: E1124 14:05:31.598038 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.598043 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.599062 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-metadata" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.599089 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" containerName="nova-metadata-log" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.600008 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.605807 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.615675 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.615808 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.653488 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-config-data\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.654028 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-logs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.654102 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.654128 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.654161 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxg42\" (UniqueName: \"kubernetes.io/projected/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-kube-api-access-hxg42\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.755755 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxg42\" (UniqueName: \"kubernetes.io/projected/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-kube-api-access-hxg42\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.755930 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-config-data\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.755987 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-logs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.756071 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.756120 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.756813 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-logs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.760733 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.761516 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-config-data\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.762371 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.783857 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxg42\" (UniqueName: \"kubernetes.io/projected/9244354d-e3fc-411a-bdc4-2d81cc2b72a0-kube-api-access-hxg42\") pod \"nova-metadata-0\" (UID: \"9244354d-e3fc-411a-bdc4-2d81cc2b72a0\") " pod="openstack/nova-metadata-0" Nov 24 14:05:31 crc kubenswrapper[4970]: I1124 14:05:31.928012 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 14:05:32 crc kubenswrapper[4970]: I1124 14:05:32.400902 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 14:05:32 crc kubenswrapper[4970]: I1124 14:05:32.548617 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9244354d-e3fc-411a-bdc4-2d81cc2b72a0","Type":"ContainerStarted","Data":"0c61008da6dbb06a13e935aef81d2e15fc6f819e3601503e6bc72ee94253a1c9"} Nov 24 14:05:32 crc kubenswrapper[4970]: I1124 14:05:32.550242 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"92b5c70e-7c5c-465f-a44c-adc0aed6c664","Type":"ContainerStarted","Data":"f262a684d1623f383c2d397f0914de835c9e3178dfc632979d6ab57c84e23047"} Nov 24 14:05:32 crc kubenswrapper[4970]: I1124 14:05:32.568514 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.56849615 podStartE2EDuration="2.56849615s" podCreationTimestamp="2025-11-24 14:05:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:32.567557036 +0000 UTC m=+1147.855314329" watchObservedRunningTime="2025-11-24 14:05:32.56849615 +0000 UTC m=+1147.856253443" Nov 24 14:05:33 crc kubenswrapper[4970]: I1124 14:05:33.485834 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaae95d4-ac49-48bf-8a37-ba8aca50eff9" path="/var/lib/kubelet/pods/eaae95d4-ac49-48bf-8a37-ba8aca50eff9/volumes" Nov 24 14:05:33 crc kubenswrapper[4970]: I1124 14:05:33.582809 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9244354d-e3fc-411a-bdc4-2d81cc2b72a0","Type":"ContainerStarted","Data":"8dbea6bf657ef227dfbc9687fb2109ef26a89be1216990c8745b556c9f40f2dc"} Nov 24 14:05:33 crc kubenswrapper[4970]: I1124 14:05:33.583600 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9244354d-e3fc-411a-bdc4-2d81cc2b72a0","Type":"ContainerStarted","Data":"3512c9177480423fafbda043c3c7dbd33f29d04dc4a365af02d92fb2f3121c6d"} Nov 24 14:05:33 crc kubenswrapper[4970]: I1124 14:05:33.619030 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.619001366 podStartE2EDuration="2.619001366s" podCreationTimestamp="2025-11-24 14:05:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:05:33.61010715 +0000 UTC m=+1148.897864463" watchObservedRunningTime="2025-11-24 14:05:33.619001366 +0000 UTC m=+1148.906758699" Nov 24 14:05:35 crc kubenswrapper[4970]: I1124 14:05:35.918029 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 14:05:36 crc kubenswrapper[4970]: I1124 14:05:36.929029 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:05:36 crc kubenswrapper[4970]: I1124 14:05:36.929098 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 14:05:37 crc kubenswrapper[4970]: I1124 14:05:37.694693 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:05:37 crc kubenswrapper[4970]: I1124 14:05:37.695880 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 14:05:38 crc kubenswrapper[4970]: I1124 14:05:38.710817 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05e7b6fd-92df-4c42-b84e-151948459084" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:38 crc kubenswrapper[4970]: I1124 14:05:38.711246 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="05e7b6fd-92df-4c42-b84e-151948459084" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:40 crc kubenswrapper[4970]: I1124 14:05:40.918312 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 14:05:40 crc kubenswrapper[4970]: I1124 14:05:40.967185 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 14:05:41 crc kubenswrapper[4970]: I1124 14:05:41.724932 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 14:05:41 crc kubenswrapper[4970]: I1124 14:05:41.929399 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:05:41 crc kubenswrapper[4970]: I1124 14:05:41.930140 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 14:05:42 crc kubenswrapper[4970]: I1124 14:05:42.435943 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 14:05:42 crc kubenswrapper[4970]: I1124 14:05:42.944787 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9244354d-e3fc-411a-bdc4-2d81cc2b72a0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:42 crc kubenswrapper[4970]: I1124 14:05:42.944787 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9244354d-e3fc-411a-bdc4-2d81cc2b72a0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.208:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.704493 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.705369 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.707154 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.742300 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.760025 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 14:05:47 crc kubenswrapper[4970]: I1124 14:05:47.772632 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 14:05:51 crc kubenswrapper[4970]: I1124 14:05:51.936507 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:05:51 crc kubenswrapper[4970]: I1124 14:05:51.939840 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 14:05:51 crc kubenswrapper[4970]: I1124 14:05:51.948402 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:05:52 crc kubenswrapper[4970]: I1124 14:05:52.823928 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 14:06:00 crc kubenswrapper[4970]: I1124 14:06:00.695728 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:02 crc kubenswrapper[4970]: I1124 14:06:02.088207 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:04 crc kubenswrapper[4970]: I1124 14:06:04.336022 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="rabbitmq" containerID="cri-o://ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f" gracePeriod=604797 Nov 24 14:06:06 crc kubenswrapper[4970]: I1124 14:06:06.272737 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="rabbitmq" containerID="cri-o://af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58" gracePeriod=604796 Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.969745 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.996803 4970 generic.go:334] "Generic (PLEG): container finished" podID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerID="ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f" exitCode=0 Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.996857 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.996866 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerDied","Data":"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f"} Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.996906 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74a8ae4a-b366-490b-aced-d4cad00248c0","Type":"ContainerDied","Data":"5b9ae162b1bdb365fef44b1e019f69922355c711a92fb7ef68d7a9add242e9ee"} Nov 24 14:06:10 crc kubenswrapper[4970]: I1124 14:06:10.996933 4970 scope.go:117] "RemoveContainer" containerID="ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.065879 4970 scope.go:117] "RemoveContainer" containerID="f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066100 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066210 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066251 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066289 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066314 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066373 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066419 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066459 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066518 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndd2p\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066536 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.066551 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins\") pod \"74a8ae4a-b366-490b-aced-d4cad00248c0\" (UID: \"74a8ae4a-b366-490b-aced-d4cad00248c0\") " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.072751 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.073388 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.074139 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.074523 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.077015 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info" (OuterVolumeSpecName: "pod-info") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.079519 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.082918 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p" (OuterVolumeSpecName: "kube-api-access-ndd2p") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "kube-api-access-ndd2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.086680 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.111320 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data" (OuterVolumeSpecName: "config-data") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.147077 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf" (OuterVolumeSpecName: "server-conf") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168676 4970 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74a8ae4a-b366-490b-aced-d4cad00248c0-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168704 4970 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74a8ae4a-b366-490b-aced-d4cad00248c0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168715 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndd2p\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-kube-api-access-ndd2p\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168724 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168733 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168755 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168764 4970 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168772 4970 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168779 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.168787 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74a8ae4a-b366-490b-aced-d4cad00248c0-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.179509 4970 scope.go:117] "RemoveContainer" containerID="ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f" Nov 24 14:06:11 crc kubenswrapper[4970]: E1124 14:06:11.179913 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f\": container with ID starting with ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f not found: ID does not exist" containerID="ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.179955 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f"} err="failed to get container status \"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f\": rpc error: code = NotFound desc = could not find container \"ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f\": container with ID starting with ee64ce1ac0172eaee39cf3d8203fe4247bddee4fdc0c7c0fa513f6fde9b0182f not found: ID does not exist" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.179983 4970 scope.go:117] "RemoveContainer" containerID="f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe" Nov 24 14:06:11 crc kubenswrapper[4970]: E1124 14:06:11.180328 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe\": container with ID starting with f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe not found: ID does not exist" containerID="f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.180353 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe"} err="failed to get container status \"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe\": rpc error: code = NotFound desc = could not find container \"f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe\": container with ID starting with f339e744ac4ee1f21c61626be2b5c71dc6417ccbce3bd5352d60d4caf85271fe not found: ID does not exist" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.188590 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.224309 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "74a8ae4a-b366-490b-aced-d4cad00248c0" (UID: "74a8ae4a-b366-490b-aced-d4cad00248c0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.270972 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.271011 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74a8ae4a-b366-490b-aced-d4cad00248c0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.339709 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.351974 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.376753 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:11 crc kubenswrapper[4970]: E1124 14:06:11.377208 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="rabbitmq" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.377230 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="rabbitmq" Nov 24 14:06:11 crc kubenswrapper[4970]: E1124 14:06:11.377251 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="setup-container" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.377262 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="setup-container" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.377460 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" containerName="rabbitmq" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.378716 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.381138 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.387322 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.387614 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.387754 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-fwx5r" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.387877 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.388051 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.388170 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.425339 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480182 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a8ae4a-b366-490b-aced-d4cad00248c0" path="/var/lib/kubelet/pods/74a8ae4a-b366-490b-aced-d4cad00248c0/volumes" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480247 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480287 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e9054df-765d-49f9-a41e-5bc3fb732691-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480308 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480444 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480508 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdgtk\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-kube-api-access-fdgtk\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480535 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480830 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480910 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480966 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.480992 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e9054df-765d-49f9-a41e-5bc3fb732691-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.481010 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.582856 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.582907 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e9054df-765d-49f9-a41e-5bc3fb732691-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.583420 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.583668 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.584237 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-config-data\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.584426 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.584848 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdgtk\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-kube-api-access-fdgtk\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.584905 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.584931 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.585062 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.585114 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.585144 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e9054df-765d-49f9-a41e-5bc3fb732691-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.585171 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.585851 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.586053 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.586825 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.587343 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6e9054df-765d-49f9-a41e-5bc3fb732691-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.590396 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6e9054df-765d-49f9-a41e-5bc3fb732691-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.590632 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.590717 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.591224 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6e9054df-765d-49f9-a41e-5bc3fb732691-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.610369 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdgtk\" (UniqueName: \"kubernetes.io/projected/6e9054df-765d-49f9-a41e-5bc3fb732691-kube-api-access-fdgtk\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.617678 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"6e9054df-765d-49f9-a41e-5bc3fb732691\") " pod="openstack/rabbitmq-server-0" Nov 24 14:06:11 crc kubenswrapper[4970]: I1124 14:06:11.698650 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 14:06:12 crc kubenswrapper[4970]: I1124 14:06:12.250933 4970 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Nov 24 14:06:12 crc kubenswrapper[4970]: I1124 14:06:12.293422 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 14:06:12 crc kubenswrapper[4970]: I1124 14:06:12.898341 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.014764 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.014819 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.014852 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.014931 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.014959 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015005 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015030 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015056 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015132 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015158 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015594 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015751 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015889 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.015310 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq6ff\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff\") pod \"94b10374-2a8e-4307-b5a2-fd9091376c12\" (UID: \"94b10374-2a8e-4307-b5a2-fd9091376c12\") " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.017248 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.017277 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.017290 4970 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.021036 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.021364 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff" (OuterVolumeSpecName: "kube-api-access-rq6ff") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "kube-api-access-rq6ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.021877 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info" (OuterVolumeSpecName: "pod-info") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.022325 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.025762 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.042558 4970 generic.go:334] "Generic (PLEG): container finished" podID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerID="af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58" exitCode=0 Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.042640 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.042676 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerDied","Data":"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58"} Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.042711 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"94b10374-2a8e-4307-b5a2-fd9091376c12","Type":"ContainerDied","Data":"81aec99ec46ae49b39341c917c65896f82d894ee50d19f5e8d7bd3ae27843d2f"} Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.042735 4970 scope.go:117] "RemoveContainer" containerID="af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.044134 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e9054df-765d-49f9-a41e-5bc3fb732691","Type":"ContainerStarted","Data":"0ebe43783161c1cff73e608bf7d2876304fa9016c54aaa51b3af2eb3305ec164"} Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.046802 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data" (OuterVolumeSpecName: "config-data") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.077323 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf" (OuterVolumeSpecName: "server-conf") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119601 4970 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94b10374-2a8e-4307-b5a2-fd9091376c12-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119634 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119646 4970 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94b10374-2a8e-4307-b5a2-fd9091376c12-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119659 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119697 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119709 4970 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94b10374-2a8e-4307-b5a2-fd9091376c12-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.119720 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq6ff\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-kube-api-access-rq6ff\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.139984 4970 scope.go:117] "RemoveContainer" containerID="0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.140893 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "94b10374-2a8e-4307-b5a2-fd9091376c12" (UID: "94b10374-2a8e-4307-b5a2-fd9091376c12"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.154865 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.166077 4970 scope.go:117] "RemoveContainer" containerID="af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58" Nov 24 14:06:13 crc kubenswrapper[4970]: E1124 14:06:13.166540 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58\": container with ID starting with af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58 not found: ID does not exist" containerID="af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.166594 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58"} err="failed to get container status \"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58\": rpc error: code = NotFound desc = could not find container \"af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58\": container with ID starting with af8fec5914cecdfeb3a9d44b1acf1be36e01e5ec3864b643d5f3d2cf3e325e58 not found: ID does not exist" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.166639 4970 scope.go:117] "RemoveContainer" containerID="0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974" Nov 24 14:06:13 crc kubenswrapper[4970]: E1124 14:06:13.169168 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974\": container with ID starting with 0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974 not found: ID does not exist" containerID="0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.169218 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974"} err="failed to get container status \"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974\": rpc error: code = NotFound desc = could not find container \"0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974\": container with ID starting with 0d79a1efaff2c5488382379c38c9fb8ed2d8e3ce972bdc72021482e3572d3974 not found: ID does not exist" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.221091 4970 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94b10374-2a8e-4307-b5a2-fd9091376c12-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.221149 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.402060 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.414801 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.437195 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:13 crc kubenswrapper[4970]: E1124 14:06:13.437891 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="setup-container" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.437943 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="setup-container" Nov 24 14:06:13 crc kubenswrapper[4970]: E1124 14:06:13.437999 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="rabbitmq" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.438020 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="rabbitmq" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.438479 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" containerName="rabbitmq" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.444271 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.462809 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.464842 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-mz7hf" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.465050 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.465464 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.466057 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.466291 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.466446 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.517990 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b10374-2a8e-4307-b5a2-fd9091376c12" path="/var/lib/kubelet/pods/94b10374-2a8e-4307-b5a2-fd9091376c12/volumes" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.518756 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526307 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526365 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4db0a4-ca88-4271-84f8-8c917057e031-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526408 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526434 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526457 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526506 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526525 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526616 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjcmm\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-kube-api-access-rjcmm\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526646 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526718 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.526738 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4db0a4-ca88-4271-84f8-8c917057e031-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627771 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627812 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4db0a4-ca88-4271-84f8-8c917057e031-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627843 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627874 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4db0a4-ca88-4271-84f8-8c917057e031-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627900 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627925 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627948 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.627981 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628000 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628032 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjcmm\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-kube-api-access-rjcmm\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628050 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628309 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628389 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.628730 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.629790 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.629919 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.630178 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9f4db0a4-ca88-4271-84f8-8c917057e031-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.631783 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9f4db0a4-ca88-4271-84f8-8c917057e031-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.642652 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjcmm\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-kube-api-access-rjcmm\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.642967 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.662726 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9f4db0a4-ca88-4271-84f8-8c917057e031-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.664104 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9f4db0a4-ca88-4271-84f8-8c917057e031-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.694149 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9f4db0a4-ca88-4271-84f8-8c917057e031\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.824224 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.884831 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.887343 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.889950 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.911229 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932021 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5njp\" (UniqueName: \"kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932059 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932103 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932130 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932164 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932181 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.932205 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:13 crc kubenswrapper[4970]: I1124 14:06:13.987321 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.011356 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d5cf5b645-zwshh"] Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.013170 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.023190 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d5cf5b645-zwshh"] Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032835 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032884 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032916 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032938 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032967 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.032984 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033000 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-svc\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033025 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033046 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-config\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033079 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033108 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033161 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5njp\" (UniqueName: \"kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033182 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.033204 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfvhl\" (UniqueName: \"kubernetes.io/projected/83916562-6f47-4b0b-8d48-6e48f0104ae6-kube-api-access-zfvhl\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.034268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.034809 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.035142 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.035148 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.035452 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.035514 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.057162 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5njp\" (UniqueName: \"kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp\") pod \"dnsmasq-dns-7f7944d86c-pcrxw\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.102287 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135315 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfvhl\" (UniqueName: \"kubernetes.io/projected/83916562-6f47-4b0b-8d48-6e48f0104ae6-kube-api-access-zfvhl\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135394 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135479 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135537 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135557 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-svc\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135607 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-config\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.135672 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.136543 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.136784 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.136929 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.136970 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-config\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.136937 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-dns-svc\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.137076 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83916562-6f47-4b0b-8d48-6e48f0104ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.156221 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfvhl\" (UniqueName: \"kubernetes.io/projected/83916562-6f47-4b0b-8d48-6e48f0104ae6-kube-api-access-zfvhl\") pod \"dnsmasq-dns-5d5cf5b645-zwshh\" (UID: \"83916562-6f47-4b0b-8d48-6e48f0104ae6\") " pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.386984 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 14:06:14 crc kubenswrapper[4970]: W1124 14:06:14.388192 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f4db0a4_ca88_4271_84f8_8c917057e031.slice/crio-151f48c30cf0ca2e8922181ed8dc1b457c517ca5cf59d84d04131af1a83bd8fd WatchSource:0}: Error finding container 151f48c30cf0ca2e8922181ed8dc1b457c517ca5cf59d84d04131af1a83bd8fd: Status 404 returned error can't find the container with id 151f48c30cf0ca2e8922181ed8dc1b457c517ca5cf59d84d04131af1a83bd8fd Nov 24 14:06:14 crc kubenswrapper[4970]: I1124 14:06:14.411864 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:15 crc kubenswrapper[4970]: I1124 14:06:15.079746 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e9054df-765d-49f9-a41e-5bc3fb732691","Type":"ContainerStarted","Data":"f489e7e9b098abb99a02bf9c6087bef2f3ec713ef8fb8d56702943a7ef7af855"} Nov 24 14:06:15 crc kubenswrapper[4970]: I1124 14:06:15.082190 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4db0a4-ca88-4271-84f8-8c917057e031","Type":"ContainerStarted","Data":"151f48c30cf0ca2e8922181ed8dc1b457c517ca5cf59d84d04131af1a83bd8fd"} Nov 24 14:06:15 crc kubenswrapper[4970]: I1124 14:06:15.203363 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d5cf5b645-zwshh"] Nov 24 14:06:15 crc kubenswrapper[4970]: I1124 14:06:15.239038 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:15 crc kubenswrapper[4970]: W1124 14:06:15.246942 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf63db9_6ee6_4e05_b6a5_f49831719a41.slice/crio-729112243cbf4f9bf194467185d71d27d370ea1307878738643bb844d78d595a WatchSource:0}: Error finding container 729112243cbf4f9bf194467185d71d27d370ea1307878738643bb844d78d595a: Status 404 returned error can't find the container with id 729112243cbf4f9bf194467185d71d27d370ea1307878738643bb844d78d595a Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.093965 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" event={"ID":"83916562-6f47-4b0b-8d48-6e48f0104ae6","Type":"ContainerStarted","Data":"afffa4f84095e338973e4b5f4f7bdf0d7494c75f23f85e4f87505076498a530d"} Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.094375 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" event={"ID":"83916562-6f47-4b0b-8d48-6e48f0104ae6","Type":"ContainerStarted","Data":"57e9e09b3c00607c07bbc50cab8f197f01a345a48c26892188059596b865bc1e"} Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.096479 4970 generic.go:334] "Generic (PLEG): container finished" podID="daf63db9-6ee6-4e05-b6a5-f49831719a41" containerID="6a8870943b89dcbb87cb1c5e1362694b6a86cbcba0f67376fe5997c624a98fc7" exitCode=0 Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.096550 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" event={"ID":"daf63db9-6ee6-4e05-b6a5-f49831719a41","Type":"ContainerDied","Data":"6a8870943b89dcbb87cb1c5e1362694b6a86cbcba0f67376fe5997c624a98fc7"} Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.096608 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" event={"ID":"daf63db9-6ee6-4e05-b6a5-f49831719a41","Type":"ContainerStarted","Data":"729112243cbf4f9bf194467185d71d27d370ea1307878738643bb844d78d595a"} Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.476377 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.590143 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.590563 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.590692 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5njp\" (UniqueName: \"kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.590776 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.590866 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.591016 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.591463 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam\") pod \"daf63db9-6ee6-4e05-b6a5-f49831719a41\" (UID: \"daf63db9-6ee6-4e05-b6a5-f49831719a41\") " Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.614895 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp" (OuterVolumeSpecName: "kube-api-access-f5njp") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "kube-api-access-f5njp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.616906 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.622054 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.625418 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config" (OuterVolumeSpecName: "config") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.626446 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.627714 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.630154 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "daf63db9-6ee6-4e05-b6a5-f49831719a41" (UID: "daf63db9-6ee6-4e05-b6a5-f49831719a41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.693962 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694003 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694013 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5njp\" (UniqueName: \"kubernetes.io/projected/daf63db9-6ee6-4e05-b6a5-f49831719a41-kube-api-access-f5njp\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694024 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694034 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694043 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:16 crc kubenswrapper[4970]: I1124 14:06:16.694051 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/daf63db9-6ee6-4e05-b6a5-f49831719a41-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.107903 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4db0a4-ca88-4271-84f8-8c917057e031","Type":"ContainerStarted","Data":"680716b30f39bda9ae6a9614bdcfb878ab5f7080d416053283226f00b6901ae0"} Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.111541 4970 generic.go:334] "Generic (PLEG): container finished" podID="83916562-6f47-4b0b-8d48-6e48f0104ae6" containerID="afffa4f84095e338973e4b5f4f7bdf0d7494c75f23f85e4f87505076498a530d" exitCode=0 Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.111632 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" event={"ID":"83916562-6f47-4b0b-8d48-6e48f0104ae6","Type":"ContainerDied","Data":"afffa4f84095e338973e4b5f4f7bdf0d7494c75f23f85e4f87505076498a530d"} Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.114178 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" event={"ID":"daf63db9-6ee6-4e05-b6a5-f49831719a41","Type":"ContainerDied","Data":"729112243cbf4f9bf194467185d71d27d370ea1307878738643bb844d78d595a"} Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.114229 4970 scope.go:117] "RemoveContainer" containerID="6a8870943b89dcbb87cb1c5e1362694b6a86cbcba0f67376fe5997c624a98fc7" Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.114319 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7944d86c-pcrxw" Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.335717 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.346742 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f7944d86c-pcrxw"] Nov 24 14:06:17 crc kubenswrapper[4970]: I1124 14:06:17.490087 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf63db9-6ee6-4e05-b6a5-f49831719a41" path="/var/lib/kubelet/pods/daf63db9-6ee6-4e05-b6a5-f49831719a41/volumes" Nov 24 14:06:18 crc kubenswrapper[4970]: I1124 14:06:18.129888 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" event={"ID":"83916562-6f47-4b0b-8d48-6e48f0104ae6","Type":"ContainerStarted","Data":"bb8fa06e56cf4aaffb8d3104717063928543a3fe921a9c3e128e7d09b7ceb556"} Nov 24 14:06:18 crc kubenswrapper[4970]: I1124 14:06:18.130027 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:18 crc kubenswrapper[4970]: I1124 14:06:18.166405 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" podStartSLOduration=5.166377164 podStartE2EDuration="5.166377164s" podCreationTimestamp="2025-11-24 14:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:06:18.163615601 +0000 UTC m=+1193.451372934" watchObservedRunningTime="2025-11-24 14:06:18.166377164 +0000 UTC m=+1193.454134497" Nov 24 14:06:24 crc kubenswrapper[4970]: I1124 14:06:24.413849 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d5cf5b645-zwshh" Nov 24 14:06:24 crc kubenswrapper[4970]: I1124 14:06:24.497851 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:06:24 crc kubenswrapper[4970]: I1124 14:06:24.498087 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="dnsmasq-dns" containerID="cri-o://024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69" gracePeriod=10 Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.054936 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205398 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205459 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lgfx\" (UniqueName: \"kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205563 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205635 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205674 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.205764 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config\") pod \"69bf2307-7427-46de-89a0-18c31d0b0c73\" (UID: \"69bf2307-7427-46de-89a0-18c31d0b0c73\") " Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.209391 4970 generic.go:334] "Generic (PLEG): container finished" podID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerID="024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69" exitCode=0 Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.209447 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.209467 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" event={"ID":"69bf2307-7427-46de-89a0-18c31d0b0c73","Type":"ContainerDied","Data":"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69"} Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.209528 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55bfb77665-7g8g8" event={"ID":"69bf2307-7427-46de-89a0-18c31d0b0c73","Type":"ContainerDied","Data":"60b466328a69432792adfb251ed837abf7869cb1490fbfaa2aa8245c55925268"} Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.209552 4970 scope.go:117] "RemoveContainer" containerID="024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.211795 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx" (OuterVolumeSpecName: "kube-api-access-5lgfx") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "kube-api-access-5lgfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.270652 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.276445 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.287742 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.292123 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.299051 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config" (OuterVolumeSpecName: "config") pod "69bf2307-7427-46de-89a0-18c31d0b0c73" (UID: "69bf2307-7427-46de-89a0-18c31d0b0c73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308235 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308289 4970 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308308 4970 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308326 4970 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308345 4970 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69bf2307-7427-46de-89a0-18c31d0b0c73-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.308361 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lgfx\" (UniqueName: \"kubernetes.io/projected/69bf2307-7427-46de-89a0-18c31d0b0c73-kube-api-access-5lgfx\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.353348 4970 scope.go:117] "RemoveContainer" containerID="449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.376020 4970 scope.go:117] "RemoveContainer" containerID="024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69" Nov 24 14:06:25 crc kubenswrapper[4970]: E1124 14:06:25.376474 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69\": container with ID starting with 024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69 not found: ID does not exist" containerID="024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.376507 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69"} err="failed to get container status \"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69\": rpc error: code = NotFound desc = could not find container \"024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69\": container with ID starting with 024de194b28c325c869d82b5975a7aa4fe2ac54c4fe9b3259b2da93819a8ba69 not found: ID does not exist" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.376528 4970 scope.go:117] "RemoveContainer" containerID="449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7" Nov 24 14:06:25 crc kubenswrapper[4970]: E1124 14:06:25.376790 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7\": container with ID starting with 449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7 not found: ID does not exist" containerID="449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.376815 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7"} err="failed to get container status \"449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7\": rpc error: code = NotFound desc = could not find container \"449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7\": container with ID starting with 449dc681ca157e0b658396f04b48ece4b4308ae6e390397401f615245180d0a7 not found: ID does not exist" Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.537141 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:06:25 crc kubenswrapper[4970]: I1124 14:06:25.548037 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55bfb77665-7g8g8"] Nov 24 14:06:27 crc kubenswrapper[4970]: I1124 14:06:27.487656 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" path="/var/lib/kubelet/pods/69bf2307-7427-46de-89a0-18c31d0b0c73/volumes" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.103322 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r"] Nov 24 14:06:33 crc kubenswrapper[4970]: E1124 14:06:33.104332 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="dnsmasq-dns" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.104350 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="dnsmasq-dns" Nov 24 14:06:33 crc kubenswrapper[4970]: E1124 14:06:33.104380 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="init" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.104389 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="init" Nov 24 14:06:33 crc kubenswrapper[4970]: E1124 14:06:33.104417 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf63db9-6ee6-4e05-b6a5-f49831719a41" containerName="init" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.104425 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf63db9-6ee6-4e05-b6a5-f49831719a41" containerName="init" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.104662 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf63db9-6ee6-4e05-b6a5-f49831719a41" containerName="init" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.104701 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="69bf2307-7427-46de-89a0-18c31d0b0c73" containerName="dnsmasq-dns" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.105405 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.111513 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.111599 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.112148 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.112455 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.115592 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r"] Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.184834 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.184910 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.184983 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwpjk\" (UniqueName: \"kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.185009 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.287318 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.287391 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwpjk\" (UniqueName: \"kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.287643 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.287713 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.294658 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.295527 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.296223 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.308233 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwpjk\" (UniqueName: \"kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:33 crc kubenswrapper[4970]: I1124 14:06:33.437951 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:34 crc kubenswrapper[4970]: I1124 14:06:34.077068 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r"] Nov 24 14:06:34 crc kubenswrapper[4970]: I1124 14:06:34.085654 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:06:34 crc kubenswrapper[4970]: I1124 14:06:34.363559 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" event={"ID":"dafb717b-e6d5-41f4-8dab-acbd1db9f563","Type":"ContainerStarted","Data":"96bf601290fc9fdddd017677121a3819992cc0485a1707a894d0fd5277ccf47c"} Nov 24 14:06:44 crc kubenswrapper[4970]: I1124 14:06:44.480850 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" event={"ID":"dafb717b-e6d5-41f4-8dab-acbd1db9f563","Type":"ContainerStarted","Data":"d6640d4efd6f9d6451d5412298320d64884aa941069b79f996b1c4064b8a10b8"} Nov 24 14:06:44 crc kubenswrapper[4970]: I1124 14:06:44.500055 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" podStartSLOduration=1.910788902 podStartE2EDuration="11.500038317s" podCreationTimestamp="2025-11-24 14:06:33 +0000 UTC" firstStartedPulling="2025-11-24 14:06:34.085390182 +0000 UTC m=+1209.373147475" lastFinishedPulling="2025-11-24 14:06:43.674639557 +0000 UTC m=+1218.962396890" observedRunningTime="2025-11-24 14:06:44.49941994 +0000 UTC m=+1219.787177243" watchObservedRunningTime="2025-11-24 14:06:44.500038317 +0000 UTC m=+1219.787795630" Nov 24 14:06:46 crc kubenswrapper[4970]: I1124 14:06:46.507311 4970 generic.go:334] "Generic (PLEG): container finished" podID="6e9054df-765d-49f9-a41e-5bc3fb732691" containerID="f489e7e9b098abb99a02bf9c6087bef2f3ec713ef8fb8d56702943a7ef7af855" exitCode=0 Nov 24 14:06:46 crc kubenswrapper[4970]: I1124 14:06:46.507412 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e9054df-765d-49f9-a41e-5bc3fb732691","Type":"ContainerDied","Data":"f489e7e9b098abb99a02bf9c6087bef2f3ec713ef8fb8d56702943a7ef7af855"} Nov 24 14:06:47 crc kubenswrapper[4970]: I1124 14:06:47.523484 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6e9054df-765d-49f9-a41e-5bc3fb732691","Type":"ContainerStarted","Data":"44e8cddd14dbe285b3833f157fd6cb52ee113de8b4ef1cb19d53dce758e9f166"} Nov 24 14:06:47 crc kubenswrapper[4970]: I1124 14:06:47.524002 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 14:06:47 crc kubenswrapper[4970]: I1124 14:06:47.553514 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.553497494 podStartE2EDuration="36.553497494s" podCreationTimestamp="2025-11-24 14:06:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:06:47.551656406 +0000 UTC m=+1222.839413739" watchObservedRunningTime="2025-11-24 14:06:47.553497494 +0000 UTC m=+1222.841254787" Nov 24 14:06:49 crc kubenswrapper[4970]: I1124 14:06:49.546035 4970 generic.go:334] "Generic (PLEG): container finished" podID="9f4db0a4-ca88-4271-84f8-8c917057e031" containerID="680716b30f39bda9ae6a9614bdcfb878ab5f7080d416053283226f00b6901ae0" exitCode=0 Nov 24 14:06:49 crc kubenswrapper[4970]: I1124 14:06:49.546133 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4db0a4-ca88-4271-84f8-8c917057e031","Type":"ContainerDied","Data":"680716b30f39bda9ae6a9614bdcfb878ab5f7080d416053283226f00b6901ae0"} Nov 24 14:06:50 crc kubenswrapper[4970]: I1124 14:06:50.560426 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9f4db0a4-ca88-4271-84f8-8c917057e031","Type":"ContainerStarted","Data":"ddf74f80c7081c23e99f12837295b06502e7239c2a16ab8b3b91190e44ff00b1"} Nov 24 14:06:50 crc kubenswrapper[4970]: I1124 14:06:50.561228 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:06:50 crc kubenswrapper[4970]: I1124 14:06:50.607129 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.607106167 podStartE2EDuration="37.607106167s" podCreationTimestamp="2025-11-24 14:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:06:50.601396526 +0000 UTC m=+1225.889153809" watchObservedRunningTime="2025-11-24 14:06:50.607106167 +0000 UTC m=+1225.894863470" Nov 24 14:06:54 crc kubenswrapper[4970]: I1124 14:06:54.602993 4970 generic.go:334] "Generic (PLEG): container finished" podID="dafb717b-e6d5-41f4-8dab-acbd1db9f563" containerID="d6640d4efd6f9d6451d5412298320d64884aa941069b79f996b1c4064b8a10b8" exitCode=0 Nov 24 14:06:54 crc kubenswrapper[4970]: I1124 14:06:54.603074 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" event={"ID":"dafb717b-e6d5-41f4-8dab-acbd1db9f563","Type":"ContainerDied","Data":"d6640d4efd6f9d6451d5412298320d64884aa941069b79f996b1c4064b8a10b8"} Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.013401 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.156709 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key\") pod \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.156763 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwpjk\" (UniqueName: \"kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk\") pod \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.156798 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle\") pod \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.156820 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory\") pod \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\" (UID: \"dafb717b-e6d5-41f4-8dab-acbd1db9f563\") " Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.163166 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "dafb717b-e6d5-41f4-8dab-acbd1db9f563" (UID: "dafb717b-e6d5-41f4-8dab-acbd1db9f563"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.172074 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk" (OuterVolumeSpecName: "kube-api-access-hwpjk") pod "dafb717b-e6d5-41f4-8dab-acbd1db9f563" (UID: "dafb717b-e6d5-41f4-8dab-acbd1db9f563"). InnerVolumeSpecName "kube-api-access-hwpjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.196985 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dafb717b-e6d5-41f4-8dab-acbd1db9f563" (UID: "dafb717b-e6d5-41f4-8dab-acbd1db9f563"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.203696 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory" (OuterVolumeSpecName: "inventory") pod "dafb717b-e6d5-41f4-8dab-acbd1db9f563" (UID: "dafb717b-e6d5-41f4-8dab-acbd1db9f563"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.258979 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.259032 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwpjk\" (UniqueName: \"kubernetes.io/projected/dafb717b-e6d5-41f4-8dab-acbd1db9f563-kube-api-access-hwpjk\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.259046 4970 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.259056 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dafb717b-e6d5-41f4-8dab-acbd1db9f563-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.624115 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" event={"ID":"dafb717b-e6d5-41f4-8dab-acbd1db9f563","Type":"ContainerDied","Data":"96bf601290fc9fdddd017677121a3819992cc0485a1707a894d0fd5277ccf47c"} Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.624459 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96bf601290fc9fdddd017677121a3819992cc0485a1707a894d0fd5277ccf47c" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.624415 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.709760 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7"] Nov 24 14:06:56 crc kubenswrapper[4970]: E1124 14:06:56.710115 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafb717b-e6d5-41f4-8dab-acbd1db9f563" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.710132 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafb717b-e6d5-41f4-8dab-acbd1db9f563" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.710351 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafb717b-e6d5-41f4-8dab-acbd1db9f563" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.711092 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.718293 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.718688 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.719286 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.729189 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7"] Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.730398 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.868552 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.868891 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.868998 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbf72\" (UniqueName: \"kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.971808 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.971900 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbf72\" (UniqueName: \"kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.972037 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.977375 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.983846 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:56 crc kubenswrapper[4970]: I1124 14:06:56.996078 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbf72\" (UniqueName: \"kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-c9km7\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:57 crc kubenswrapper[4970]: I1124 14:06:57.035681 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:06:57 crc kubenswrapper[4970]: I1124 14:06:57.635782 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7"] Nov 24 14:06:58 crc kubenswrapper[4970]: I1124 14:06:58.647048 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" event={"ID":"f6e35ba6-d32d-4204-a731-7f0e24802382","Type":"ContainerStarted","Data":"a77a3d70b55cd8e35d3b7b0e33de45ead811f54c4bde1bd7a6f3b651438b289e"} Nov 24 14:06:58 crc kubenswrapper[4970]: I1124 14:06:58.647798 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" event={"ID":"f6e35ba6-d32d-4204-a731-7f0e24802382","Type":"ContainerStarted","Data":"c39f4e1f7bcba520fff876bcc54c5107e121aea6cf4babcf8551457fa42b891e"} Nov 24 14:06:58 crc kubenswrapper[4970]: I1124 14:06:58.668482 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" podStartSLOduration=2.125056667 podStartE2EDuration="2.668456991s" podCreationTimestamp="2025-11-24 14:06:56 +0000 UTC" firstStartedPulling="2025-11-24 14:06:57.646636505 +0000 UTC m=+1232.934393798" lastFinishedPulling="2025-11-24 14:06:58.190036829 +0000 UTC m=+1233.477794122" observedRunningTime="2025-11-24 14:06:58.663772466 +0000 UTC m=+1233.951529779" watchObservedRunningTime="2025-11-24 14:06:58.668456991 +0000 UTC m=+1233.956214304" Nov 24 14:07:01 crc kubenswrapper[4970]: I1124 14:07:01.680180 4970 generic.go:334] "Generic (PLEG): container finished" podID="f6e35ba6-d32d-4204-a731-7f0e24802382" containerID="a77a3d70b55cd8e35d3b7b0e33de45ead811f54c4bde1bd7a6f3b651438b289e" exitCode=0 Nov 24 14:07:01 crc kubenswrapper[4970]: I1124 14:07:01.680338 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" event={"ID":"f6e35ba6-d32d-4204-a731-7f0e24802382","Type":"ContainerDied","Data":"a77a3d70b55cd8e35d3b7b0e33de45ead811f54c4bde1bd7a6f3b651438b289e"} Nov 24 14:07:01 crc kubenswrapper[4970]: I1124 14:07:01.703737 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.155593 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.297691 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory\") pod \"f6e35ba6-d32d-4204-a731-7f0e24802382\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.297752 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key\") pod \"f6e35ba6-d32d-4204-a731-7f0e24802382\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.298703 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbf72\" (UniqueName: \"kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72\") pod \"f6e35ba6-d32d-4204-a731-7f0e24802382\" (UID: \"f6e35ba6-d32d-4204-a731-7f0e24802382\") " Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.306064 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72" (OuterVolumeSpecName: "kube-api-access-xbf72") pod "f6e35ba6-d32d-4204-a731-7f0e24802382" (UID: "f6e35ba6-d32d-4204-a731-7f0e24802382"). InnerVolumeSpecName "kube-api-access-xbf72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.329443 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory" (OuterVolumeSpecName: "inventory") pod "f6e35ba6-d32d-4204-a731-7f0e24802382" (UID: "f6e35ba6-d32d-4204-a731-7f0e24802382"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.339566 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f6e35ba6-d32d-4204-a731-7f0e24802382" (UID: "f6e35ba6-d32d-4204-a731-7f0e24802382"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.402186 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbf72\" (UniqueName: \"kubernetes.io/projected/f6e35ba6-d32d-4204-a731-7f0e24802382-kube-api-access-xbf72\") on node \"crc\" DevicePath \"\"" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.402227 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.402240 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f6e35ba6-d32d-4204-a731-7f0e24802382-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.699935 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" event={"ID":"f6e35ba6-d32d-4204-a731-7f0e24802382","Type":"ContainerDied","Data":"c39f4e1f7bcba520fff876bcc54c5107e121aea6cf4babcf8551457fa42b891e"} Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.699974 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c39f4e1f7bcba520fff876bcc54c5107e121aea6cf4babcf8551457fa42b891e" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.700037 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-c9km7" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.772643 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv"] Nov 24 14:07:03 crc kubenswrapper[4970]: E1124 14:07:03.773014 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e35ba6-d32d-4204-a731-7f0e24802382" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.773031 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e35ba6-d32d-4204-a731-7f0e24802382" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.773214 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e35ba6-d32d-4204-a731-7f0e24802382" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.773809 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.775498 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.776181 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.776324 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.776937 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.793387 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv"] Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.827644 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.914794 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.914846 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.914904 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:03 crc kubenswrapper[4970]: I1124 14:07:03.914967 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4q4v\" (UniqueName: \"kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.015993 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.016081 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4q4v\" (UniqueName: \"kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.016172 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.016194 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.020292 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.020319 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.020666 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.036067 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4q4v\" (UniqueName: \"kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.089536 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.659794 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv"] Nov 24 14:07:04 crc kubenswrapper[4970]: W1124 14:07:04.675465 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9700095_d8fb_496e_981d_de1e30ad7e8a.slice/crio-27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74 WatchSource:0}: Error finding container 27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74: Status 404 returned error can't find the container with id 27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74 Nov 24 14:07:04 crc kubenswrapper[4970]: I1124 14:07:04.709233 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" event={"ID":"b9700095-d8fb-496e-981d-de1e30ad7e8a","Type":"ContainerStarted","Data":"27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74"} Nov 24 14:07:05 crc kubenswrapper[4970]: I1124 14:07:05.719094 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" event={"ID":"b9700095-d8fb-496e-981d-de1e30ad7e8a","Type":"ContainerStarted","Data":"e1fe4779b0f9838f81245632b90ec8a9fb9092c7eafd3bfe052b1477aba37c06"} Nov 24 14:07:05 crc kubenswrapper[4970]: I1124 14:07:05.736762 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" podStartSLOduration=2.298851853 podStartE2EDuration="2.736739511s" podCreationTimestamp="2025-11-24 14:07:03 +0000 UTC" firstStartedPulling="2025-11-24 14:07:04.678780847 +0000 UTC m=+1239.966538140" lastFinishedPulling="2025-11-24 14:07:05.116668485 +0000 UTC m=+1240.404425798" observedRunningTime="2025-11-24 14:07:05.734115601 +0000 UTC m=+1241.021872894" watchObservedRunningTime="2025-11-24 14:07:05.736739511 +0000 UTC m=+1241.024496804" Nov 24 14:07:11 crc kubenswrapper[4970]: I1124 14:07:11.204839 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:07:11 crc kubenswrapper[4970]: I1124 14:07:11.205690 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:07:41 crc kubenswrapper[4970]: I1124 14:07:41.203733 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:07:41 crc kubenswrapper[4970]: I1124 14:07:41.204328 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:07:44 crc kubenswrapper[4970]: I1124 14:07:44.408294 4970 scope.go:117] "RemoveContainer" containerID="efee003b2d183170e97231f87e1b0975182e3d3f192ac172e0de50d21de2c70b" Nov 24 14:07:44 crc kubenswrapper[4970]: I1124 14:07:44.447446 4970 scope.go:117] "RemoveContainer" containerID="023a6dcfc6a9b3fe97be4a430419e54f0851c25121d4d43603098a17c4d21c3b" Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.203905 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.204494 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.204541 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.205391 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.205459 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57" gracePeriod=600 Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.465699 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57" exitCode=0 Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.466018 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57"} Nov 24 14:08:11 crc kubenswrapper[4970]: I1124 14:08:11.466055 4970 scope.go:117] "RemoveContainer" containerID="6f40f7e031fb1e58c5777cffe6a105a8691f418bd2d11cdfc1447f31eb4b8f3e" Nov 24 14:08:12 crc kubenswrapper[4970]: I1124 14:08:12.479379 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270"} Nov 24 14:08:44 crc kubenswrapper[4970]: I1124 14:08:44.555744 4970 scope.go:117] "RemoveContainer" containerID="8ffdb7fd15ddb66c3c218b48044bfd5556055b39b03a73c6674eb36af583b0a2" Nov 24 14:09:44 crc kubenswrapper[4970]: I1124 14:09:44.656903 4970 scope.go:117] "RemoveContainer" containerID="c48df9554a6eac5c7de114da9d4e9d3d33153f4f11c4a691480a39cf0cd01b1d" Nov 24 14:09:44 crc kubenswrapper[4970]: I1124 14:09:44.687714 4970 scope.go:117] "RemoveContainer" containerID="25d030299d57e4ef5ade75a2cba774943a4477dcb0e7d4d461d152708684cf0f" Nov 24 14:10:00 crc kubenswrapper[4970]: I1124 14:10:00.615521 4970 generic.go:334] "Generic (PLEG): container finished" podID="b9700095-d8fb-496e-981d-de1e30ad7e8a" containerID="e1fe4779b0f9838f81245632b90ec8a9fb9092c7eafd3bfe052b1477aba37c06" exitCode=0 Nov 24 14:10:00 crc kubenswrapper[4970]: I1124 14:10:00.615637 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" event={"ID":"b9700095-d8fb-496e-981d-de1e30ad7e8a","Type":"ContainerDied","Data":"e1fe4779b0f9838f81245632b90ec8a9fb9092c7eafd3bfe052b1477aba37c06"} Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.038309 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.085124 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle\") pod \"b9700095-d8fb-496e-981d-de1e30ad7e8a\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.085200 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4q4v\" (UniqueName: \"kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v\") pod \"b9700095-d8fb-496e-981d-de1e30ad7e8a\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.085262 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory\") pod \"b9700095-d8fb-496e-981d-de1e30ad7e8a\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.085368 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key\") pod \"b9700095-d8fb-496e-981d-de1e30ad7e8a\" (UID: \"b9700095-d8fb-496e-981d-de1e30ad7e8a\") " Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.090999 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b9700095-d8fb-496e-981d-de1e30ad7e8a" (UID: "b9700095-d8fb-496e-981d-de1e30ad7e8a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.091063 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v" (OuterVolumeSpecName: "kube-api-access-k4q4v") pod "b9700095-d8fb-496e-981d-de1e30ad7e8a" (UID: "b9700095-d8fb-496e-981d-de1e30ad7e8a"). InnerVolumeSpecName "kube-api-access-k4q4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.111877 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b9700095-d8fb-496e-981d-de1e30ad7e8a" (UID: "b9700095-d8fb-496e-981d-de1e30ad7e8a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.112823 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory" (OuterVolumeSpecName: "inventory") pod "b9700095-d8fb-496e-981d-de1e30ad7e8a" (UID: "b9700095-d8fb-496e-981d-de1e30ad7e8a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.187316 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.187354 4970 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.187368 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4q4v\" (UniqueName: \"kubernetes.io/projected/b9700095-d8fb-496e-981d-de1e30ad7e8a-kube-api-access-k4q4v\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.187379 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b9700095-d8fb-496e-981d-de1e30ad7e8a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.647550 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" event={"ID":"b9700095-d8fb-496e-981d-de1e30ad7e8a","Type":"ContainerDied","Data":"27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74"} Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.647936 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27c3988d7591e230e5e6d26571bb3914e91a5dd5a01ddc3a805ac73c82594a74" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.647669 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.722679 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8"] Nov 24 14:10:02 crc kubenswrapper[4970]: E1124 14:10:02.723130 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9700095-d8fb-496e-981d-de1e30ad7e8a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.723149 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9700095-d8fb-496e-981d-de1e30ad7e8a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.723367 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9700095-d8fb-496e-981d-de1e30ad7e8a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.724129 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.726554 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.726660 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.726733 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.728284 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.736977 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8"] Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.797702 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.797776 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.797892 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wrd7\" (UniqueName: \"kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.899856 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wrd7\" (UniqueName: \"kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.899966 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.900018 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.903377 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.903833 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:02 crc kubenswrapper[4970]: I1124 14:10:02.915829 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wrd7\" (UniqueName: \"kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:03 crc kubenswrapper[4970]: I1124 14:10:03.081807 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:10:03 crc kubenswrapper[4970]: I1124 14:10:03.636172 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8"] Nov 24 14:10:03 crc kubenswrapper[4970]: I1124 14:10:03.659935 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" event={"ID":"bdec1f38-8d93-4298-ad14-872b58cff4e9","Type":"ContainerStarted","Data":"577b2db1560ed451e826f31e96a79faaaf605e014de16825b528f5fce50fc86e"} Nov 24 14:10:04 crc kubenswrapper[4970]: I1124 14:10:04.670670 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" event={"ID":"bdec1f38-8d93-4298-ad14-872b58cff4e9","Type":"ContainerStarted","Data":"49d6c10bcf37e38f3e741ebe61ba70c4fc06f1b1e1a987e68588477a250816d7"} Nov 24 14:10:04 crc kubenswrapper[4970]: I1124 14:10:04.696654 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" podStartSLOduration=2.163277049 podStartE2EDuration="2.696636826s" podCreationTimestamp="2025-11-24 14:10:02 +0000 UTC" firstStartedPulling="2025-11-24 14:10:03.64563099 +0000 UTC m=+1418.933388293" lastFinishedPulling="2025-11-24 14:10:04.178990767 +0000 UTC m=+1419.466748070" observedRunningTime="2025-11-24 14:10:04.690201461 +0000 UTC m=+1419.977958754" watchObservedRunningTime="2025-11-24 14:10:04.696636826 +0000 UTC m=+1419.984394119" Nov 24 14:10:11 crc kubenswrapper[4970]: I1124 14:10:11.204605 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:10:11 crc kubenswrapper[4970]: I1124 14:10:11.205043 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.081334 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.086199 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.104503 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.180841 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pbdf\" (UniqueName: \"kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.180969 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.181248 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.282853 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.282952 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pbdf\" (UniqueName: \"kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.282998 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.283448 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.283427 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.307495 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pbdf\" (UniqueName: \"kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf\") pod \"certified-operators-jglcz\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.418219 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:21 crc kubenswrapper[4970]: I1124 14:10:21.936174 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:22 crc kubenswrapper[4970]: I1124 14:10:22.853896 4970 generic.go:334] "Generic (PLEG): container finished" podID="a831125d-2c7c-43c3-8156-56c03f54b574" containerID="05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5" exitCode=0 Nov 24 14:10:22 crc kubenswrapper[4970]: I1124 14:10:22.854021 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerDied","Data":"05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5"} Nov 24 14:10:22 crc kubenswrapper[4970]: I1124 14:10:22.854284 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerStarted","Data":"dab9e0450715a105aeafd36d94f862f1d3186a184c2d2459141a03cd219efd08"} Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.430039 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.435624 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.455088 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.583627 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxg5\" (UniqueName: \"kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.583713 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.583753 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.685716 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.685764 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.685957 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxg5\" (UniqueName: \"kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.686310 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.686428 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.714066 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxg5\" (UniqueName: \"kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5\") pod \"redhat-operators-hk5w7\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:25 crc kubenswrapper[4970]: I1124 14:10:25.792657 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:26 crc kubenswrapper[4970]: I1124 14:10:26.280487 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:10:26 crc kubenswrapper[4970]: I1124 14:10:26.905544 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerStarted","Data":"e1525a3f60445daca4e8ba0554bff007d77525760f4d0a2b246ad7050947d48a"} Nov 24 14:10:28 crc kubenswrapper[4970]: I1124 14:10:28.930129 4970 generic.go:334] "Generic (PLEG): container finished" podID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerID="06e530c66f85f0747bbfb32721e9b5a87992b83388d36f67926573fba693b0ae" exitCode=0 Nov 24 14:10:28 crc kubenswrapper[4970]: I1124 14:10:28.930226 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerDied","Data":"06e530c66f85f0747bbfb32721e9b5a87992b83388d36f67926573fba693b0ae"} Nov 24 14:10:29 crc kubenswrapper[4970]: I1124 14:10:29.948731 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerStarted","Data":"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e"} Nov 24 14:10:30 crc kubenswrapper[4970]: I1124 14:10:30.962230 4970 generic.go:334] "Generic (PLEG): container finished" podID="a831125d-2c7c-43c3-8156-56c03f54b574" containerID="dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e" exitCode=0 Nov 24 14:10:30 crc kubenswrapper[4970]: I1124 14:10:30.962281 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerDied","Data":"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e"} Nov 24 14:10:34 crc kubenswrapper[4970]: I1124 14:10:34.000467 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerStarted","Data":"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353"} Nov 24 14:10:34 crc kubenswrapper[4970]: I1124 14:10:34.005807 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerStarted","Data":"61618aec90053c020a383f53d62e916e97dd79ce0658925b77410a2a756ae419"} Nov 24 14:10:34 crc kubenswrapper[4970]: I1124 14:10:34.028324 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jglcz" podStartSLOduration=2.6499147389999997 podStartE2EDuration="13.028304029s" podCreationTimestamp="2025-11-24 14:10:21 +0000 UTC" firstStartedPulling="2025-11-24 14:10:22.858829359 +0000 UTC m=+1438.146586692" lastFinishedPulling="2025-11-24 14:10:33.237218659 +0000 UTC m=+1448.524975982" observedRunningTime="2025-11-24 14:10:34.025626489 +0000 UTC m=+1449.313383792" watchObservedRunningTime="2025-11-24 14:10:34.028304029 +0000 UTC m=+1449.316061322" Nov 24 14:10:41 crc kubenswrapper[4970]: I1124 14:10:41.203929 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:10:41 crc kubenswrapper[4970]: I1124 14:10:41.204830 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:10:41 crc kubenswrapper[4970]: I1124 14:10:41.419558 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:41 crc kubenswrapper[4970]: I1124 14:10:41.419634 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:41 crc kubenswrapper[4970]: I1124 14:10:41.480702 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:42 crc kubenswrapper[4970]: I1124 14:10:42.205247 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:42 crc kubenswrapper[4970]: I1124 14:10:42.268224 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.169223 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jglcz" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="registry-server" containerID="cri-o://81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353" gracePeriod=2 Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.630976 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.760201 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content\") pod \"a831125d-2c7c-43c3-8156-56c03f54b574\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.760416 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pbdf\" (UniqueName: \"kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf\") pod \"a831125d-2c7c-43c3-8156-56c03f54b574\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.760476 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities\") pod \"a831125d-2c7c-43c3-8156-56c03f54b574\" (UID: \"a831125d-2c7c-43c3-8156-56c03f54b574\") " Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.761161 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities" (OuterVolumeSpecName: "utilities") pod "a831125d-2c7c-43c3-8156-56c03f54b574" (UID: "a831125d-2c7c-43c3-8156-56c03f54b574"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.766134 4970 scope.go:117] "RemoveContainer" containerID="26e36d6d8d834f11c4fc7136e5c9bb504ac9e0413b4ead7d36b8eb75e0a031fd" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.777802 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf" (OuterVolumeSpecName: "kube-api-access-4pbdf") pod "a831125d-2c7c-43c3-8156-56c03f54b574" (UID: "a831125d-2c7c-43c3-8156-56c03f54b574"). InnerVolumeSpecName "kube-api-access-4pbdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.805506 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a831125d-2c7c-43c3-8156-56c03f54b574" (UID: "a831125d-2c7c-43c3-8156-56c03f54b574"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.862562 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pbdf\" (UniqueName: \"kubernetes.io/projected/a831125d-2c7c-43c3-8156-56c03f54b574-kube-api-access-4pbdf\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.862621 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:44 crc kubenswrapper[4970]: I1124 14:10:44.862633 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a831125d-2c7c-43c3-8156-56c03f54b574-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.180178 4970 generic.go:334] "Generic (PLEG): container finished" podID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerID="61618aec90053c020a383f53d62e916e97dd79ce0658925b77410a2a756ae419" exitCode=0 Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.180267 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerDied","Data":"61618aec90053c020a383f53d62e916e97dd79ce0658925b77410a2a756ae419"} Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.186331 4970 generic.go:334] "Generic (PLEG): container finished" podID="a831125d-2c7c-43c3-8156-56c03f54b574" containerID="81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353" exitCode=0 Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.186395 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jglcz" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.186415 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerDied","Data":"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353"} Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.186801 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jglcz" event={"ID":"a831125d-2c7c-43c3-8156-56c03f54b574","Type":"ContainerDied","Data":"dab9e0450715a105aeafd36d94f862f1d3186a184c2d2459141a03cd219efd08"} Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.186842 4970 scope.go:117] "RemoveContainer" containerID="81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.230377 4970 scope.go:117] "RemoveContainer" containerID="dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.236357 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.248797 4970 scope.go:117] "RemoveContainer" containerID="05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.251176 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jglcz"] Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.298618 4970 scope.go:117] "RemoveContainer" containerID="81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353" Nov 24 14:10:45 crc kubenswrapper[4970]: E1124 14:10:45.299058 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353\": container with ID starting with 81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353 not found: ID does not exist" containerID="81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.299093 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353"} err="failed to get container status \"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353\": rpc error: code = NotFound desc = could not find container \"81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353\": container with ID starting with 81866f8465d4f29259a9765fb1040d5deb6aca657cb922a1eee5137c00ba4353 not found: ID does not exist" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.299118 4970 scope.go:117] "RemoveContainer" containerID="dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e" Nov 24 14:10:45 crc kubenswrapper[4970]: E1124 14:10:45.299385 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e\": container with ID starting with dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e not found: ID does not exist" containerID="dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.299424 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e"} err="failed to get container status \"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e\": rpc error: code = NotFound desc = could not find container \"dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e\": container with ID starting with dd2a4585338a47b242945bdc1d86e5bb0e7e93e2765d2dd9dfc0dccae410687e not found: ID does not exist" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.299449 4970 scope.go:117] "RemoveContainer" containerID="05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5" Nov 24 14:10:45 crc kubenswrapper[4970]: E1124 14:10:45.299891 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5\": container with ID starting with 05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5 not found: ID does not exist" containerID="05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.299912 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5"} err="failed to get container status \"05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5\": rpc error: code = NotFound desc = could not find container \"05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5\": container with ID starting with 05cf5c379c476bba3503bc4451695ae810eac5048371d84a1cb83d95d24820b5 not found: ID does not exist" Nov 24 14:10:45 crc kubenswrapper[4970]: I1124 14:10:45.487818 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" path="/var/lib/kubelet/pods/a831125d-2c7c-43c3-8156-56c03f54b574/volumes" Nov 24 14:10:46 crc kubenswrapper[4970]: I1124 14:10:46.207703 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerStarted","Data":"2454be05bb3a43170c82a837d224171f87eff0e64effccc7f1cfd99fb321dd04"} Nov 24 14:10:55 crc kubenswrapper[4970]: I1124 14:10:55.792979 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:55 crc kubenswrapper[4970]: I1124 14:10:55.794797 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:10:56 crc kubenswrapper[4970]: I1124 14:10:56.845219 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hk5w7" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="registry-server" probeResult="failure" output=< Nov 24 14:10:56 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 14:10:56 crc kubenswrapper[4970]: > Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.645856 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hk5w7" podStartSLOduration=19.890225843 podStartE2EDuration="36.645833416s" podCreationTimestamp="2025-11-24 14:10:25 +0000 UTC" firstStartedPulling="2025-11-24 14:10:28.931734452 +0000 UTC m=+1444.219491745" lastFinishedPulling="2025-11-24 14:10:45.687342015 +0000 UTC m=+1460.975099318" observedRunningTime="2025-11-24 14:10:46.230610296 +0000 UTC m=+1461.518367589" watchObservedRunningTime="2025-11-24 14:11:01.645833416 +0000 UTC m=+1476.933590709" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.651150 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:01 crc kubenswrapper[4970]: E1124 14:11:01.651626 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="extract-utilities" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.651648 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="extract-utilities" Nov 24 14:11:01 crc kubenswrapper[4970]: E1124 14:11:01.651675 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="registry-server" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.651685 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="registry-server" Nov 24 14:11:01 crc kubenswrapper[4970]: E1124 14:11:01.651699 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="extract-content" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.651706 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="extract-content" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.652027 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a831125d-2c7c-43c3-8156-56c03f54b574" containerName="registry-server" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.653743 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.665043 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.797311 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zqb\" (UniqueName: \"kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.797633 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.797697 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.899989 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zqb\" (UniqueName: \"kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.900053 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.900103 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.900614 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.900940 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.934665 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zqb\" (UniqueName: \"kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb\") pod \"redhat-marketplace-l477c\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:01 crc kubenswrapper[4970]: I1124 14:11:01.976447 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:02 crc kubenswrapper[4970]: I1124 14:11:02.441165 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:03 crc kubenswrapper[4970]: I1124 14:11:03.362706 4970 generic.go:334] "Generic (PLEG): container finished" podID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerID="98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d" exitCode=0 Nov 24 14:11:03 crc kubenswrapper[4970]: I1124 14:11:03.362769 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerDied","Data":"98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d"} Nov 24 14:11:03 crc kubenswrapper[4970]: I1124 14:11:03.363113 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerStarted","Data":"3c51a6559d0fecf6799e4fc8d4e5e01a0bf7e67b8faa791e0de93245398678a2"} Nov 24 14:11:04 crc kubenswrapper[4970]: I1124 14:11:04.397295 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerStarted","Data":"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72"} Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.409168 4970 generic.go:334] "Generic (PLEG): container finished" podID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerID="074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72" exitCode=0 Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.409272 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerDied","Data":"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72"} Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.409534 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerStarted","Data":"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0"} Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.435236 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l477c" podStartSLOduration=2.993902859 podStartE2EDuration="4.435214695s" podCreationTimestamp="2025-11-24 14:11:01 +0000 UTC" firstStartedPulling="2025-11-24 14:11:03.369054317 +0000 UTC m=+1478.656811630" lastFinishedPulling="2025-11-24 14:11:04.810366143 +0000 UTC m=+1480.098123466" observedRunningTime="2025-11-24 14:11:05.426291566 +0000 UTC m=+1480.714048869" watchObservedRunningTime="2025-11-24 14:11:05.435214695 +0000 UTC m=+1480.722971988" Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.839909 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:11:05 crc kubenswrapper[4970]: I1124 14:11:05.900150 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.020609 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.021129 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hk5w7" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="registry-server" containerID="cri-o://2454be05bb3a43170c82a837d224171f87eff0e64effccc7f1cfd99fb321dd04" gracePeriod=2 Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.437706 4970 generic.go:334] "Generic (PLEG): container finished" podID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerID="2454be05bb3a43170c82a837d224171f87eff0e64effccc7f1cfd99fb321dd04" exitCode=0 Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.437788 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerDied","Data":"2454be05bb3a43170c82a837d224171f87eff0e64effccc7f1cfd99fb321dd04"} Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.438189 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hk5w7" event={"ID":"2535d24a-c9e6-4d76-9407-d3628c17f52f","Type":"ContainerDied","Data":"e1525a3f60445daca4e8ba0554bff007d77525760f4d0a2b246ad7050947d48a"} Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.438217 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1525a3f60445daca4e8ba0554bff007d77525760f4d0a2b246ad7050947d48a" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.489409 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.629568 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmxg5\" (UniqueName: \"kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5\") pod \"2535d24a-c9e6-4d76-9407-d3628c17f52f\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.629643 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content\") pod \"2535d24a-c9e6-4d76-9407-d3628c17f52f\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.629685 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities\") pod \"2535d24a-c9e6-4d76-9407-d3628c17f52f\" (UID: \"2535d24a-c9e6-4d76-9407-d3628c17f52f\") " Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.630505 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities" (OuterVolumeSpecName: "utilities") pod "2535d24a-c9e6-4d76-9407-d3628c17f52f" (UID: "2535d24a-c9e6-4d76-9407-d3628c17f52f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.635270 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5" (OuterVolumeSpecName: "kube-api-access-dmxg5") pod "2535d24a-c9e6-4d76-9407-d3628c17f52f" (UID: "2535d24a-c9e6-4d76-9407-d3628c17f52f"). InnerVolumeSpecName "kube-api-access-dmxg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.732746 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmxg5\" (UniqueName: \"kubernetes.io/projected/2535d24a-c9e6-4d76-9407-d3628c17f52f-kube-api-access-dmxg5\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.732738 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2535d24a-c9e6-4d76-9407-d3628c17f52f" (UID: "2535d24a-c9e6-4d76-9407-d3628c17f52f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.732781 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:08 crc kubenswrapper[4970]: I1124 14:11:08.834212 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2535d24a-c9e6-4d76-9407-d3628c17f52f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:09 crc kubenswrapper[4970]: I1124 14:11:09.450795 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hk5w7" Nov 24 14:11:09 crc kubenswrapper[4970]: I1124 14:11:09.513296 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:11:09 crc kubenswrapper[4970]: I1124 14:11:09.526060 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hk5w7"] Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.203930 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.204214 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.204255 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.204912 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.204955 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" gracePeriod=600 Nov 24 14:11:11 crc kubenswrapper[4970]: E1124 14:11:11.357622 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.473515 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" exitCode=0 Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.483645 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" path="/var/lib/kubelet/pods/2535d24a-c9e6-4d76-9407-d3628c17f52f/volumes" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.484674 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270"} Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.484815 4970 scope.go:117] "RemoveContainer" containerID="0b9ca7fe93072c90c317323214d5177c999803f2d302d20f63ac180ac1902e57" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.485724 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:11:11 crc kubenswrapper[4970]: E1124 14:11:11.486243 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.977600 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:11 crc kubenswrapper[4970]: I1124 14:11:11.977939 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:12 crc kubenswrapper[4970]: I1124 14:11:12.027671 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:12 crc kubenswrapper[4970]: I1124 14:11:12.600148 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:13 crc kubenswrapper[4970]: I1124 14:11:13.226269 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:14 crc kubenswrapper[4970]: I1124 14:11:14.514111 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l477c" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="registry-server" containerID="cri-o://71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0" gracePeriod=2 Nov 24 14:11:14 crc kubenswrapper[4970]: I1124 14:11:14.925088 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.065058 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content\") pod \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.065189 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zqb\" (UniqueName: \"kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb\") pod \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.065227 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities\") pod \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\" (UID: \"81e87814-38a1-4d8a-8f92-22ab5e81a93e\") " Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.066302 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities" (OuterVolumeSpecName: "utilities") pod "81e87814-38a1-4d8a-8f92-22ab5e81a93e" (UID: "81e87814-38a1-4d8a-8f92-22ab5e81a93e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.076072 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb" (OuterVolumeSpecName: "kube-api-access-99zqb") pod "81e87814-38a1-4d8a-8f92-22ab5e81a93e" (UID: "81e87814-38a1-4d8a-8f92-22ab5e81a93e"). InnerVolumeSpecName "kube-api-access-99zqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.084361 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81e87814-38a1-4d8a-8f92-22ab5e81a93e" (UID: "81e87814-38a1-4d8a-8f92-22ab5e81a93e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.166859 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99zqb\" (UniqueName: \"kubernetes.io/projected/81e87814-38a1-4d8a-8f92-22ab5e81a93e-kube-api-access-99zqb\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.167105 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.167192 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81e87814-38a1-4d8a-8f92-22ab5e81a93e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.528710 4970 generic.go:334] "Generic (PLEG): container finished" podID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerID="71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0" exitCode=0 Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.528833 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l477c" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.528862 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerDied","Data":"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0"} Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.530457 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l477c" event={"ID":"81e87814-38a1-4d8a-8f92-22ab5e81a93e","Type":"ContainerDied","Data":"3c51a6559d0fecf6799e4fc8d4e5e01a0bf7e67b8faa791e0de93245398678a2"} Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.530517 4970 scope.go:117] "RemoveContainer" containerID="71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.567270 4970 scope.go:117] "RemoveContainer" containerID="074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.568116 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.582989 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l477c"] Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.594337 4970 scope.go:117] "RemoveContainer" containerID="98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.657720 4970 scope.go:117] "RemoveContainer" containerID="71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0" Nov 24 14:11:15 crc kubenswrapper[4970]: E1124 14:11:15.658241 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0\": container with ID starting with 71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0 not found: ID does not exist" containerID="71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.658282 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0"} err="failed to get container status \"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0\": rpc error: code = NotFound desc = could not find container \"71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0\": container with ID starting with 71122123071c1df7c996f6126c90e427a504166717f0c8fe07ef046249f877b0 not found: ID does not exist" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.658305 4970 scope.go:117] "RemoveContainer" containerID="074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72" Nov 24 14:11:15 crc kubenswrapper[4970]: E1124 14:11:15.658781 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72\": container with ID starting with 074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72 not found: ID does not exist" containerID="074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.658803 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72"} err="failed to get container status \"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72\": rpc error: code = NotFound desc = could not find container \"074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72\": container with ID starting with 074dfcb0ac99fb1a5fbcc3a6fdf09ad66984be4a7d37f76a8d0b66acbcb71e72 not found: ID does not exist" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.658818 4970 scope.go:117] "RemoveContainer" containerID="98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d" Nov 24 14:11:15 crc kubenswrapper[4970]: E1124 14:11:15.659148 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d\": container with ID starting with 98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d not found: ID does not exist" containerID="98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d" Nov 24 14:11:15 crc kubenswrapper[4970]: I1124 14:11:15.659169 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d"} err="failed to get container status \"98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d\": rpc error: code = NotFound desc = could not find container \"98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d\": container with ID starting with 98f913905e831aed941b90f7fcd3ccaf0849474c98d826ec5682fe58e7e6771d not found: ID does not exist" Nov 24 14:11:17 crc kubenswrapper[4970]: I1124 14:11:17.484529 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" path="/var/lib/kubelet/pods/81e87814-38a1-4d8a-8f92-22ab5e81a93e/volumes" Nov 24 14:11:23 crc kubenswrapper[4970]: I1124 14:11:23.471494 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:11:23 crc kubenswrapper[4970]: E1124 14:11:23.472293 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:11:24 crc kubenswrapper[4970]: I1124 14:11:24.039384 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cksvv"] Nov 24 14:11:24 crc kubenswrapper[4970]: I1124 14:11:24.048184 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-3696-account-create-z754x"] Nov 24 14:11:24 crc kubenswrapper[4970]: I1124 14:11:24.056517 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-3696-account-create-z754x"] Nov 24 14:11:24 crc kubenswrapper[4970]: I1124 14:11:24.064170 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cksvv"] Nov 24 14:11:25 crc kubenswrapper[4970]: I1124 14:11:25.487764 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bbc61d9-f25b-47c6-879e-db93d0e543f2" path="/var/lib/kubelet/pods/6bbc61d9-f25b-47c6-879e-db93d0e543f2/volumes" Nov 24 14:11:25 crc kubenswrapper[4970]: I1124 14:11:25.489162 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cef88f7b-658e-4d42-b264-20532c99aa42" path="/var/lib/kubelet/pods/cef88f7b-658e-4d42-b264-20532c99aa42/volumes" Nov 24 14:11:28 crc kubenswrapper[4970]: I1124 14:11:28.038310 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-ptwhn"] Nov 24 14:11:28 crc kubenswrapper[4970]: I1124 14:11:28.058022 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-33f1-account-create-92lmd"] Nov 24 14:11:28 crc kubenswrapper[4970]: I1124 14:11:28.066118 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-ptwhn"] Nov 24 14:11:28 crc kubenswrapper[4970]: I1124 14:11:28.074784 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-33f1-account-create-92lmd"] Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.032195 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cmgxd"] Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.040351 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9297-account-create-xkv78"] Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.049245 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cmgxd"] Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.056384 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9297-account-create-xkv78"] Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.480025 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="279dafe8-9eae-4665-a2f2-781688a11b9b" path="/var/lib/kubelet/pods/279dafe8-9eae-4665-a2f2-781688a11b9b/volumes" Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.481051 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d73b83-954e-4da1-ae25-f975d3687eeb" path="/var/lib/kubelet/pods/d8d73b83-954e-4da1-ae25-f975d3687eeb/volumes" Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.481734 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee934db6-19ef-4f00-b20a-498b1c229b4b" path="/var/lib/kubelet/pods/ee934db6-19ef-4f00-b20a-498b1c229b4b/volumes" Nov 24 14:11:29 crc kubenswrapper[4970]: I1124 14:11:29.482601 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6377db5-fe8e-404b-98f2-7ad39a001b80" path="/var/lib/kubelet/pods/f6377db5-fe8e-404b-98f2-7ad39a001b80/volumes" Nov 24 14:11:34 crc kubenswrapper[4970]: I1124 14:11:34.471153 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:11:34 crc kubenswrapper[4970]: E1124 14:11:34.472029 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:11:44 crc kubenswrapper[4970]: I1124 14:11:44.809747 4970 generic.go:334] "Generic (PLEG): container finished" podID="bdec1f38-8d93-4298-ad14-872b58cff4e9" containerID="49d6c10bcf37e38f3e741ebe61ba70c4fc06f1b1e1a987e68588477a250816d7" exitCode=0 Nov 24 14:11:44 crc kubenswrapper[4970]: I1124 14:11:44.809850 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" event={"ID":"bdec1f38-8d93-4298-ad14-872b58cff4e9","Type":"ContainerDied","Data":"49d6c10bcf37e38f3e741ebe61ba70c4fc06f1b1e1a987e68588477a250816d7"} Nov 24 14:11:44 crc kubenswrapper[4970]: I1124 14:11:44.877726 4970 scope.go:117] "RemoveContainer" containerID="43ab091a8e02ef1ed64d265cc863ea547c2be3f2d27d487b4427947eeb33f2ad" Nov 24 14:11:44 crc kubenswrapper[4970]: I1124 14:11:44.910483 4970 scope.go:117] "RemoveContainer" containerID="f96b6ea125f374eb8b3d0a56057b4d010823242f039d476d69f650c04b3453d1" Nov 24 14:11:44 crc kubenswrapper[4970]: I1124 14:11:44.975051 4970 scope.go:117] "RemoveContainer" containerID="ec359ee4bd5a0b507b92900e21efc01299987149a7222c5df7c2c8ff2a9d4a0d" Nov 24 14:11:45 crc kubenswrapper[4970]: I1124 14:11:45.014939 4970 scope.go:117] "RemoveContainer" containerID="4a516ad9356643f201a8149cee73fe0c58824c178a6cc8930fe663d976a71627" Nov 24 14:11:45 crc kubenswrapper[4970]: I1124 14:11:45.051985 4970 scope.go:117] "RemoveContainer" containerID="ace05e2c1c0446e587ac307db01dc573ec554abcdd8b3aa28c25de520a608914" Nov 24 14:11:45 crc kubenswrapper[4970]: I1124 14:11:45.089503 4970 scope.go:117] "RemoveContainer" containerID="90ed9af6d1858625b5b6043f8203a0ca6ef5a44e746ecaa6bb884b98e65a5e48" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.261176 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.367265 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key\") pod \"bdec1f38-8d93-4298-ad14-872b58cff4e9\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.367323 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wrd7\" (UniqueName: \"kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7\") pod \"bdec1f38-8d93-4298-ad14-872b58cff4e9\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.367432 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory\") pod \"bdec1f38-8d93-4298-ad14-872b58cff4e9\" (UID: \"bdec1f38-8d93-4298-ad14-872b58cff4e9\") " Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.375036 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7" (OuterVolumeSpecName: "kube-api-access-5wrd7") pod "bdec1f38-8d93-4298-ad14-872b58cff4e9" (UID: "bdec1f38-8d93-4298-ad14-872b58cff4e9"). InnerVolumeSpecName "kube-api-access-5wrd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.408546 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory" (OuterVolumeSpecName: "inventory") pod "bdec1f38-8d93-4298-ad14-872b58cff4e9" (UID: "bdec1f38-8d93-4298-ad14-872b58cff4e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.429285 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bdec1f38-8d93-4298-ad14-872b58cff4e9" (UID: "bdec1f38-8d93-4298-ad14-872b58cff4e9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.470117 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wrd7\" (UniqueName: \"kubernetes.io/projected/bdec1f38-8d93-4298-ad14-872b58cff4e9-kube-api-access-5wrd7\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.470157 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.470174 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bdec1f38-8d93-4298-ad14-872b58cff4e9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.834174 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" event={"ID":"bdec1f38-8d93-4298-ad14-872b58cff4e9","Type":"ContainerDied","Data":"577b2db1560ed451e826f31e96a79faaaf605e014de16825b528f5fce50fc86e"} Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.834522 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="577b2db1560ed451e826f31e96a79faaaf605e014de16825b528f5fce50fc86e" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.834229 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.954998 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl"] Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955402 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="extract-content" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955421 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="extract-content" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955434 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdec1f38-8d93-4298-ad14-872b58cff4e9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955442 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdec1f38-8d93-4298-ad14-872b58cff4e9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955470 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="extract-content" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955476 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="extract-content" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955488 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955495 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955504 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955509 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955518 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="extract-utilities" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955524 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="extract-utilities" Nov 24 14:11:46 crc kubenswrapper[4970]: E1124 14:11:46.955535 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="extract-utilities" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955541 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="extract-utilities" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955751 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e87814-38a1-4d8a-8f92-22ab5e81a93e" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955765 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdec1f38-8d93-4298-ad14-872b58cff4e9" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.955778 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="2535d24a-c9e6-4d76-9407-d3628c17f52f" containerName="registry-server" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.956339 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.958171 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.958171 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.958338 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.959432 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:11:46 crc kubenswrapper[4970]: I1124 14:11:46.974382 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl"] Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.031315 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-fpfvv"] Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.039380 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-fpfvv"] Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.081292 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.081368 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.081414 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnxw5\" (UniqueName: \"kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.183122 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.183252 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.183792 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnxw5\" (UniqueName: \"kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.188877 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.189102 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.201123 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnxw5\" (UniqueName: \"kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-s49zl\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.291173 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.491195 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:11:47 crc kubenswrapper[4970]: E1124 14:11:47.491815 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.506277 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ec76e6-76f2-4454-803d-5b5cb58c37d9" path="/var/lib/kubelet/pods/c9ec76e6-76f2-4454-803d-5b5cb58c37d9/volumes" Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.916202 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl"] Nov 24 14:11:47 crc kubenswrapper[4970]: W1124 14:11:47.928165 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34ffc82b_5c7e_4995_8dab_f09ed6e1c2ad.slice/crio-5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f WatchSource:0}: Error finding container 5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f: Status 404 returned error can't find the container with id 5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f Nov 24 14:11:47 crc kubenswrapper[4970]: I1124 14:11:47.930870 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:11:48 crc kubenswrapper[4970]: I1124 14:11:48.857917 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" event={"ID":"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad","Type":"ContainerStarted","Data":"07c0a314449ee35bc71f29456b1b0ecca33dba361fac7eb1dd5cbbd58005ea37"} Nov 24 14:11:48 crc kubenswrapper[4970]: I1124 14:11:48.858234 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" event={"ID":"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad","Type":"ContainerStarted","Data":"5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f"} Nov 24 14:11:48 crc kubenswrapper[4970]: I1124 14:11:48.890828 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" podStartSLOduration=2.492098316 podStartE2EDuration="2.890804754s" podCreationTimestamp="2025-11-24 14:11:46 +0000 UTC" firstStartedPulling="2025-11-24 14:11:47.930678678 +0000 UTC m=+1523.218435971" lastFinishedPulling="2025-11-24 14:11:48.329385116 +0000 UTC m=+1523.617142409" observedRunningTime="2025-11-24 14:11:48.883261148 +0000 UTC m=+1524.171018461" watchObservedRunningTime="2025-11-24 14:11:48.890804754 +0000 UTC m=+1524.178562067" Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.049677 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5205-account-create-2qjmw"] Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.065676 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6189-account-create-5fm2f"] Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.077892 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5205-account-create-2qjmw"] Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.086974 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6189-account-create-5fm2f"] Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.095090 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-k696g"] Nov 24 14:11:56 crc kubenswrapper[4970]: I1124 14:11:56.113728 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-k696g"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.031710 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-z5stn"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.040804 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8446-account-create-kqdzx"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.048390 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-z5stn"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.055674 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-wkwfz"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.062344 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-wkwfz"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.068755 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8446-account-create-kqdzx"] Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.481037 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489" path="/var/lib/kubelet/pods/7d1ae5b1-853f-4c6c-9a9f-d2e6db92a489/volumes" Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.481631 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a055662-deb7-4f6a-bef1-a69e71f155e0" path="/var/lib/kubelet/pods/8a055662-deb7-4f6a-bef1-a69e71f155e0/volumes" Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.482140 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf9ba03-18af-4c11-ad00-d6e26ef65c8f" path="/var/lib/kubelet/pods/9cf9ba03-18af-4c11-ad00-d6e26ef65c8f/volumes" Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.482856 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94b8e19-9edb-4a82-95a9-5717afa5ae6d" path="/var/lib/kubelet/pods/c94b8e19-9edb-4a82-95a9-5717afa5ae6d/volumes" Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.483925 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77cc176-a7dd-4510-8f10-a5cc227ca06f" path="/var/lib/kubelet/pods/d77cc176-a7dd-4510-8f10-a5cc227ca06f/volumes" Nov 24 14:11:57 crc kubenswrapper[4970]: I1124 14:11:57.484443 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd54ba95-cb56-4068-90c9-06cfd72111c4" path="/var/lib/kubelet/pods/dd54ba95-cb56-4068-90c9-06cfd72111c4/volumes" Nov 24 14:11:58 crc kubenswrapper[4970]: I1124 14:11:58.471203 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:11:58 crc kubenswrapper[4970]: E1124 14:11:58.472051 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:12:11 crc kubenswrapper[4970]: I1124 14:12:11.038469 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c6tvk"] Nov 24 14:12:11 crc kubenswrapper[4970]: I1124 14:12:11.050223 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c6tvk"] Nov 24 14:12:11 crc kubenswrapper[4970]: I1124 14:12:11.486491 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145d1a69-5b39-4ec0-8544-3034bded1955" path="/var/lib/kubelet/pods/145d1a69-5b39-4ec0-8544-3034bded1955/volumes" Nov 24 14:12:12 crc kubenswrapper[4970]: I1124 14:12:12.470549 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:12:12 crc kubenswrapper[4970]: E1124 14:12:12.471430 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:12:25 crc kubenswrapper[4970]: I1124 14:12:25.481351 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:12:25 crc kubenswrapper[4970]: E1124 14:12:25.482947 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:12:38 crc kubenswrapper[4970]: I1124 14:12:38.470610 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:12:38 crc kubenswrapper[4970]: E1124 14:12:38.471396 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.284247 4970 scope.go:117] "RemoveContainer" containerID="33c6643207a43500f6e9e52f3b50294535ed8997fc0dc80ceb6bc37b295d350a" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.313010 4970 scope.go:117] "RemoveContainer" containerID="0590b4bc1a00e2b45b16494ec48cc4f6d6cc43c062db9ba88c2a7a53bf105265" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.379953 4970 scope.go:117] "RemoveContainer" containerID="31b20bb8ad269e9ebd09a8f86d9da47dd7aa7f4dad9109e7e2e63c54f61df4f6" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.410993 4970 scope.go:117] "RemoveContainer" containerID="e95f90e4f466ba059b65cb676c994ce48521fabee21932369b74146fb40c82c1" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.469355 4970 scope.go:117] "RemoveContainer" containerID="67129bc7ccb47df1a286fb0ff31884279c3c06c3697633e07d8bc857a1b389ba" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.504102 4970 scope.go:117] "RemoveContainer" containerID="994745c33b0ef29b802f527e53366b3bdb6621cc0946896116b5d409c5119325" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.545413 4970 scope.go:117] "RemoveContainer" containerID="dcce86608a5d568912915abfef47037ffee042a6dccceb3621416d2e2deafd02" Nov 24 14:12:45 crc kubenswrapper[4970]: I1124 14:12:45.605210 4970 scope.go:117] "RemoveContainer" containerID="9f8f5954bc3a8e93e9bbb1263f0b358632b019b48093d84865515af5bd21dbbb" Nov 24 14:12:46 crc kubenswrapper[4970]: I1124 14:12:46.041599 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jnshr"] Nov 24 14:12:46 crc kubenswrapper[4970]: I1124 14:12:46.054419 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jnshr"] Nov 24 14:12:47 crc kubenswrapper[4970]: I1124 14:12:47.514952 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9908e3-7a2c-4649-af58-9f4d24c5f06c" path="/var/lib/kubelet/pods/6d9908e3-7a2c-4649-af58-9f4d24c5f06c/volumes" Nov 24 14:12:51 crc kubenswrapper[4970]: I1124 14:12:51.046708 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-7x8kz"] Nov 24 14:12:51 crc kubenswrapper[4970]: I1124 14:12:51.057325 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-7x8kz"] Nov 24 14:12:51 crc kubenswrapper[4970]: I1124 14:12:51.483611 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d29428-7adf-4a22-b656-36c9b931c8d4" path="/var/lib/kubelet/pods/75d29428-7adf-4a22-b656-36c9b931c8d4/volumes" Nov 24 14:12:52 crc kubenswrapper[4970]: I1124 14:12:52.470826 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:12:52 crc kubenswrapper[4970]: E1124 14:12:52.471535 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:12:59 crc kubenswrapper[4970]: I1124 14:12:59.601112 4970 generic.go:334] "Generic (PLEG): container finished" podID="34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" containerID="07c0a314449ee35bc71f29456b1b0ecca33dba361fac7eb1dd5cbbd58005ea37" exitCode=0 Nov 24 14:12:59 crc kubenswrapper[4970]: I1124 14:12:59.601278 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" event={"ID":"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad","Type":"ContainerDied","Data":"07c0a314449ee35bc71f29456b1b0ecca33dba361fac7eb1dd5cbbd58005ea37"} Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.072379 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.187712 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory\") pod \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.187793 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key\") pod \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.187900 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnxw5\" (UniqueName: \"kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5\") pod \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\" (UID: \"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad\") " Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.203521 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5" (OuterVolumeSpecName: "kube-api-access-qnxw5") pod "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" (UID: "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad"). InnerVolumeSpecName "kube-api-access-qnxw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.219963 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" (UID: "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.220919 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory" (OuterVolumeSpecName: "inventory") pod "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" (UID: "34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.290834 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnxw5\" (UniqueName: \"kubernetes.io/projected/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-kube-api-access-qnxw5\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.290890 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.290914 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.622127 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" event={"ID":"34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad","Type":"ContainerDied","Data":"5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f"} Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.622180 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-s49zl" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.622191 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e1a0dce5f30281c9f2c54febb414ed6aefa514e07bd4cff156910bd46345e7f" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.719239 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w"] Nov 24 14:13:01 crc kubenswrapper[4970]: E1124 14:13:01.719718 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.719745 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.720182 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.721017 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.725940 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.725999 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.726140 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.726440 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.734202 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w"] Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.798970 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.799154 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.799345 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hgd7\" (UniqueName: \"kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.901301 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.901761 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hgd7\" (UniqueName: \"kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.901952 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.906986 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.915208 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:01 crc kubenswrapper[4970]: I1124 14:13:01.919645 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hgd7\" (UniqueName: \"kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:02 crc kubenswrapper[4970]: I1124 14:13:02.041807 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:02 crc kubenswrapper[4970]: I1124 14:13:02.609701 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w"] Nov 24 14:13:02 crc kubenswrapper[4970]: I1124 14:13:02.629640 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" event={"ID":"4665b179-69e1-41a9-a621-5681262a765a","Type":"ContainerStarted","Data":"d9d505659422b66d19b339762898c3b2668ff6b354e9402a0e464c5274bf433b"} Nov 24 14:13:03 crc kubenswrapper[4970]: I1124 14:13:03.644646 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" event={"ID":"4665b179-69e1-41a9-a621-5681262a765a","Type":"ContainerStarted","Data":"528372e36c2ac98e096dae9fba9ae854e9b3dd12d6ddfb8ef60773833fe01127"} Nov 24 14:13:03 crc kubenswrapper[4970]: I1124 14:13:03.665675 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" podStartSLOduration=2.208323864 podStartE2EDuration="2.665653097s" podCreationTimestamp="2025-11-24 14:13:01 +0000 UTC" firstStartedPulling="2025-11-24 14:13:02.61376465 +0000 UTC m=+1597.901521943" lastFinishedPulling="2025-11-24 14:13:03.071093883 +0000 UTC m=+1598.358851176" observedRunningTime="2025-11-24 14:13:03.661982433 +0000 UTC m=+1598.949739746" watchObservedRunningTime="2025-11-24 14:13:03.665653097 +0000 UTC m=+1598.953410390" Nov 24 14:13:05 crc kubenswrapper[4970]: I1124 14:13:05.028427 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-6j5rx"] Nov 24 14:13:05 crc kubenswrapper[4970]: I1124 14:13:05.036783 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-6j5rx"] Nov 24 14:13:05 crc kubenswrapper[4970]: I1124 14:13:05.489469 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:13:05 crc kubenswrapper[4970]: E1124 14:13:05.489978 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:13:05 crc kubenswrapper[4970]: I1124 14:13:05.498256 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="461c95e6-1aaf-44cc-b090-430a4379e671" path="/var/lib/kubelet/pods/461c95e6-1aaf-44cc-b090-430a4379e671/volumes" Nov 24 14:13:07 crc kubenswrapper[4970]: I1124 14:13:07.027448 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-nmb4b"] Nov 24 14:13:07 crc kubenswrapper[4970]: I1124 14:13:07.037159 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-nmb4b"] Nov 24 14:13:07 crc kubenswrapper[4970]: I1124 14:13:07.484523 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48c12423-0ed8-4c46-bfa2-249b9c67eb14" path="/var/lib/kubelet/pods/48c12423-0ed8-4c46-bfa2-249b9c67eb14/volumes" Nov 24 14:13:08 crc kubenswrapper[4970]: I1124 14:13:08.725130 4970 generic.go:334] "Generic (PLEG): container finished" podID="4665b179-69e1-41a9-a621-5681262a765a" containerID="528372e36c2ac98e096dae9fba9ae854e9b3dd12d6ddfb8ef60773833fe01127" exitCode=0 Nov 24 14:13:08 crc kubenswrapper[4970]: I1124 14:13:08.725224 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" event={"ID":"4665b179-69e1-41a9-a621-5681262a765a","Type":"ContainerDied","Data":"528372e36c2ac98e096dae9fba9ae854e9b3dd12d6ddfb8ef60773833fe01127"} Nov 24 14:13:09 crc kubenswrapper[4970]: I1124 14:13:09.039514 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-dt7wf"] Nov 24 14:13:09 crc kubenswrapper[4970]: I1124 14:13:09.045658 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-dt7wf"] Nov 24 14:13:09 crc kubenswrapper[4970]: I1124 14:13:09.481789 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85ae910-cbec-47ff-9f7e-bba2545cbcae" path="/var/lib/kubelet/pods/f85ae910-cbec-47ff-9f7e-bba2545cbcae/volumes" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.184508 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.264231 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key\") pod \"4665b179-69e1-41a9-a621-5681262a765a\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.264299 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hgd7\" (UniqueName: \"kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7\") pod \"4665b179-69e1-41a9-a621-5681262a765a\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.264396 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory\") pod \"4665b179-69e1-41a9-a621-5681262a765a\" (UID: \"4665b179-69e1-41a9-a621-5681262a765a\") " Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.275748 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7" (OuterVolumeSpecName: "kube-api-access-2hgd7") pod "4665b179-69e1-41a9-a621-5681262a765a" (UID: "4665b179-69e1-41a9-a621-5681262a765a"). InnerVolumeSpecName "kube-api-access-2hgd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.294525 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4665b179-69e1-41a9-a621-5681262a765a" (UID: "4665b179-69e1-41a9-a621-5681262a765a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.301629 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory" (OuterVolumeSpecName: "inventory") pod "4665b179-69e1-41a9-a621-5681262a765a" (UID: "4665b179-69e1-41a9-a621-5681262a765a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.365641 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.365677 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hgd7\" (UniqueName: \"kubernetes.io/projected/4665b179-69e1-41a9-a621-5681262a765a-kube-api-access-2hgd7\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.365710 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4665b179-69e1-41a9-a621-5681262a765a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.770309 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" event={"ID":"4665b179-69e1-41a9-a621-5681262a765a","Type":"ContainerDied","Data":"d9d505659422b66d19b339762898c3b2668ff6b354e9402a0e464c5274bf433b"} Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.770352 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9d505659422b66d19b339762898c3b2668ff6b354e9402a0e464c5274bf433b" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.770438 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.849866 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs"] Nov 24 14:13:10 crc kubenswrapper[4970]: E1124 14:13:10.850456 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4665b179-69e1-41a9-a621-5681262a765a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.850507 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="4665b179-69e1-41a9-a621-5681262a765a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.850994 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="4665b179-69e1-41a9-a621-5681262a765a" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.852097 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.856468 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.856734 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.857160 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.857185 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.859566 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs"] Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.975634 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.976387 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:10 crc kubenswrapper[4970]: I1124 14:13:10.976471 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n4tx\" (UniqueName: \"kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.078716 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.079084 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n4tx\" (UniqueName: \"kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.079259 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.085045 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.088121 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.104719 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n4tx\" (UniqueName: \"kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fctzs\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.175145 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.735180 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs"] Nov 24 14:13:11 crc kubenswrapper[4970]: I1124 14:13:11.779124 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" event={"ID":"50a618ec-6a8b-4234-8000-253f692b204a","Type":"ContainerStarted","Data":"a7b44b0e5b67e59bc863ec27006f0f1374bde4237e951a92d51f267537375677"} Nov 24 14:13:12 crc kubenswrapper[4970]: I1124 14:13:12.790055 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" event={"ID":"50a618ec-6a8b-4234-8000-253f692b204a","Type":"ContainerStarted","Data":"72d67ad8294d063e0a2a59b4570aa5bd7ca1a1b90f92080a6528a0bccd645787"} Nov 24 14:13:19 crc kubenswrapper[4970]: I1124 14:13:19.470803 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:13:19 crc kubenswrapper[4970]: E1124 14:13:19.471540 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.119469 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" podStartSLOduration=12.692618518 podStartE2EDuration="13.119432086s" podCreationTimestamp="2025-11-24 14:13:10 +0000 UTC" firstStartedPulling="2025-11-24 14:13:11.73728752 +0000 UTC m=+1607.025044813" lastFinishedPulling="2025-11-24 14:13:12.164101048 +0000 UTC m=+1607.451858381" observedRunningTime="2025-11-24 14:13:12.807530831 +0000 UTC m=+1608.095288124" watchObservedRunningTime="2025-11-24 14:13:23.119432086 +0000 UTC m=+1618.407189429" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.132170 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.134660 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.151239 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.314938 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.315102 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbjlx\" (UniqueName: \"kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.315467 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.416859 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.416941 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.417027 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbjlx\" (UniqueName: \"kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.417527 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.417617 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.449621 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbjlx\" (UniqueName: \"kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx\") pod \"community-operators-bnws5\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:23 crc kubenswrapper[4970]: I1124 14:13:23.458379 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:24 crc kubenswrapper[4970]: I1124 14:13:24.021345 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:13:24 crc kubenswrapper[4970]: W1124 14:13:24.026056 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod962a33e9_70d8_463f_b70b_c0427b287b5a.slice/crio-b367b218058f81ddbfe5d8430bc6f4bf807fcb4eba5c2a4bbc85c88790f31379 WatchSource:0}: Error finding container b367b218058f81ddbfe5d8430bc6f4bf807fcb4eba5c2a4bbc85c88790f31379: Status 404 returned error can't find the container with id b367b218058f81ddbfe5d8430bc6f4bf807fcb4eba5c2a4bbc85c88790f31379 Nov 24 14:13:24 crc kubenswrapper[4970]: I1124 14:13:24.915514 4970 generic.go:334] "Generic (PLEG): container finished" podID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerID="1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a" exitCode=0 Nov 24 14:13:24 crc kubenswrapper[4970]: I1124 14:13:24.915569 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerDied","Data":"1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a"} Nov 24 14:13:24 crc kubenswrapper[4970]: I1124 14:13:24.915882 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerStarted","Data":"b367b218058f81ddbfe5d8430bc6f4bf807fcb4eba5c2a4bbc85c88790f31379"} Nov 24 14:13:28 crc kubenswrapper[4970]: I1124 14:13:28.963921 4970 generic.go:334] "Generic (PLEG): container finished" podID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerID="d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd" exitCode=0 Nov 24 14:13:28 crc kubenswrapper[4970]: I1124 14:13:28.963972 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerDied","Data":"d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd"} Nov 24 14:13:29 crc kubenswrapper[4970]: I1124 14:13:29.976699 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerStarted","Data":"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234"} Nov 24 14:13:29 crc kubenswrapper[4970]: I1124 14:13:29.997925 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bnws5" podStartSLOduration=2.457012135 podStartE2EDuration="6.997905879s" podCreationTimestamp="2025-11-24 14:13:23 +0000 UTC" firstStartedPulling="2025-11-24 14:13:24.917198539 +0000 UTC m=+1620.204955832" lastFinishedPulling="2025-11-24 14:13:29.458092283 +0000 UTC m=+1624.745849576" observedRunningTime="2025-11-24 14:13:29.993850304 +0000 UTC m=+1625.281607607" watchObservedRunningTime="2025-11-24 14:13:29.997905879 +0000 UTC m=+1625.285663192" Nov 24 14:13:30 crc kubenswrapper[4970]: I1124 14:13:30.470181 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:13:30 crc kubenswrapper[4970]: E1124 14:13:30.470423 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:13:33 crc kubenswrapper[4970]: I1124 14:13:33.459163 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:33 crc kubenswrapper[4970]: I1124 14:13:33.459557 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:33 crc kubenswrapper[4970]: I1124 14:13:33.531620 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:34 crc kubenswrapper[4970]: I1124 14:13:34.061220 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:13:34 crc kubenswrapper[4970]: I1124 14:13:34.135248 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:13:34 crc kubenswrapper[4970]: I1124 14:13:34.166970 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 14:13:34 crc kubenswrapper[4970]: I1124 14:13:34.167253 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xwtxh" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="registry-server" containerID="cri-o://66e8c6276601fd762d7293cc6127dce3f778bc2bb0de3c6104b446c5a16f90c6" gracePeriod=2 Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.026458 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerID="66e8c6276601fd762d7293cc6127dce3f778bc2bb0de3c6104b446c5a16f90c6" exitCode=0 Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.026540 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerDied","Data":"66e8c6276601fd762d7293cc6127dce3f778bc2bb0de3c6104b446c5a16f90c6"} Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.837252 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.949173 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities\") pod \"5ff82462-7b38-4888-af1f-8d5eb8629c58\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.949314 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content\") pod \"5ff82462-7b38-4888-af1f-8d5eb8629c58\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.949343 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4wzh\" (UniqueName: \"kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh\") pod \"5ff82462-7b38-4888-af1f-8d5eb8629c58\" (UID: \"5ff82462-7b38-4888-af1f-8d5eb8629c58\") " Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.950152 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities" (OuterVolumeSpecName: "utilities") pod "5ff82462-7b38-4888-af1f-8d5eb8629c58" (UID: "5ff82462-7b38-4888-af1f-8d5eb8629c58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.954555 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh" (OuterVolumeSpecName: "kube-api-access-d4wzh") pod "5ff82462-7b38-4888-af1f-8d5eb8629c58" (UID: "5ff82462-7b38-4888-af1f-8d5eb8629c58"). InnerVolumeSpecName "kube-api-access-d4wzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:13:35 crc kubenswrapper[4970]: I1124 14:13:35.991366 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ff82462-7b38-4888-af1f-8d5eb8629c58" (UID: "5ff82462-7b38-4888-af1f-8d5eb8629c58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.039300 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xwtxh" event={"ID":"5ff82462-7b38-4888-af1f-8d5eb8629c58","Type":"ContainerDied","Data":"5aeaac152f7a43bd795f5c39e073b1c905738e7144a24637962065594ed584bb"} Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.039344 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xwtxh" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.039449 4970 scope.go:117] "RemoveContainer" containerID="66e8c6276601fd762d7293cc6127dce3f778bc2bb0de3c6104b446c5a16f90c6" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.051715 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.051744 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4wzh\" (UniqueName: \"kubernetes.io/projected/5ff82462-7b38-4888-af1f-8d5eb8629c58-kube-api-access-d4wzh\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.051754 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ff82462-7b38-4888-af1f-8d5eb8629c58-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.065552 4970 scope.go:117] "RemoveContainer" containerID="6689f03422c6a1dd3ea5001899e6af473996e5d53336134f25b16d306b9dee2c" Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.099013 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.107666 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xwtxh"] Nov 24 14:13:36 crc kubenswrapper[4970]: I1124 14:13:36.121866 4970 scope.go:117] "RemoveContainer" containerID="6e0214bae0969f87cd8dd43cdc6c88ea6fe79774c1eac911d6e1941016de6aea" Nov 24 14:13:37 crc kubenswrapper[4970]: I1124 14:13:37.485017 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" path="/var/lib/kubelet/pods/5ff82462-7b38-4888-af1f-8d5eb8629c58/volumes" Nov 24 14:13:44 crc kubenswrapper[4970]: I1124 14:13:44.035921 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-053e-account-create-2hclc"] Nov 24 14:13:44 crc kubenswrapper[4970]: I1124 14:13:44.045859 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-053e-account-create-2hclc"] Nov 24 14:13:44 crc kubenswrapper[4970]: I1124 14:13:44.470351 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:13:44 crc kubenswrapper[4970]: E1124 14:13:44.470677 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.040246 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9849-account-create-nkhkx"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.048870 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-tfqlt"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.057531 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-84e1-account-create-f27tz"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.065484 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9849-account-create-nkhkx"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.073045 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hbxpk"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.079386 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-84e1-account-create-f27tz"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.085178 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b5b2b"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.095635 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b5b2b"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.110977 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hbxpk"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.114904 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-tfqlt"] Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.488061 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2def0468-8470-4258-b55a-b640a71e248c" path="/var/lib/kubelet/pods/2def0468-8470-4258-b55a-b640a71e248c/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.489339 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37faa061-9cb6-4364-9bf1-7cee945a3859" path="/var/lib/kubelet/pods/37faa061-9cb6-4364-9bf1-7cee945a3859/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.490247 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5419ff1f-e0e4-453a-97e1-a3d5651077c7" path="/var/lib/kubelet/pods/5419ff1f-e0e4-453a-97e1-a3d5651077c7/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.491118 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b055400-56b2-41b3-8ecd-2f4235608a51" path="/var/lib/kubelet/pods/8b055400-56b2-41b3-8ecd-2f4235608a51/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.492706 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf2cf6d-0109-4c10-bc97-0efba40e772f" path="/var/lib/kubelet/pods/8bf2cf6d-0109-4c10-bc97-0efba40e772f/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.493558 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5f4e454-4f77-46f3-8a28-feb38fc462e6" path="/var/lib/kubelet/pods/d5f4e454-4f77-46f3-8a28-feb38fc462e6/volumes" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.793355 4970 scope.go:117] "RemoveContainer" containerID="7086366a4a9317eb1c139ef54322b201ed90d9007bfa270c3b24f2292892c37f" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.828271 4970 scope.go:117] "RemoveContainer" containerID="8c686f47662db039f8847333c14d316dd2516b1f6d03ef268cc37f5cd11be59b" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.923526 4970 scope.go:117] "RemoveContainer" containerID="996234b67812d8b08e1437234a253f4f1349e9a9d2dfbf0921407f9c13a70412" Nov 24 14:13:45 crc kubenswrapper[4970]: I1124 14:13:45.956289 4970 scope.go:117] "RemoveContainer" containerID="a274e436d6160821373d877d910076c422db0eda9a63de18d0b30e4acecef05d" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.027207 4970 scope.go:117] "RemoveContainer" containerID="c5d334c1a2f85008f81f76bd9aa00ec9c8379feeadf60e05f9dfb1cebdc0e0fd" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.061074 4970 scope.go:117] "RemoveContainer" containerID="43d8a1437b02f626bcd750f61e27e686098703af06c0141791dd0bf25ac70fea" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.118359 4970 scope.go:117] "RemoveContainer" containerID="91b7c1a2ff2869740ebfe758ca0357f2e6452f87f228a6b36f63b7b4d6893352" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.147730 4970 scope.go:117] "RemoveContainer" containerID="86867005583fbb6f73a0238863e598f48c3d0a36437cb1fe5458a206f6642b5a" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.176422 4970 scope.go:117] "RemoveContainer" containerID="5d0e351c82fb3f11bec7386180873a71cde3882a0bed300d3ef74bdd147e48cb" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.200853 4970 scope.go:117] "RemoveContainer" containerID="b41eee2584d4780cb068ef7889d33e01cc2429fe66f27fc9ce058e04290353c9" Nov 24 14:13:46 crc kubenswrapper[4970]: I1124 14:13:46.224376 4970 scope.go:117] "RemoveContainer" containerID="ccbf580b445ad9bf3c460a99f20117ec6ca78bb0de719c03ced2441378335df8" Nov 24 14:13:49 crc kubenswrapper[4970]: I1124 14:13:49.168657 4970 generic.go:334] "Generic (PLEG): container finished" podID="50a618ec-6a8b-4234-8000-253f692b204a" containerID="72d67ad8294d063e0a2a59b4570aa5bd7ca1a1b90f92080a6528a0bccd645787" exitCode=0 Nov 24 14:13:49 crc kubenswrapper[4970]: I1124 14:13:49.168838 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" event={"ID":"50a618ec-6a8b-4234-8000-253f692b204a","Type":"ContainerDied","Data":"72d67ad8294d063e0a2a59b4570aa5bd7ca1a1b90f92080a6528a0bccd645787"} Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.666939 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.749301 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n4tx\" (UniqueName: \"kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx\") pod \"50a618ec-6a8b-4234-8000-253f692b204a\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.749550 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key\") pod \"50a618ec-6a8b-4234-8000-253f692b204a\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.749708 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory\") pod \"50a618ec-6a8b-4234-8000-253f692b204a\" (UID: \"50a618ec-6a8b-4234-8000-253f692b204a\") " Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.754671 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx" (OuterVolumeSpecName: "kube-api-access-2n4tx") pod "50a618ec-6a8b-4234-8000-253f692b204a" (UID: "50a618ec-6a8b-4234-8000-253f692b204a"). InnerVolumeSpecName "kube-api-access-2n4tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.782932 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50a618ec-6a8b-4234-8000-253f692b204a" (UID: "50a618ec-6a8b-4234-8000-253f692b204a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.783356 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory" (OuterVolumeSpecName: "inventory") pod "50a618ec-6a8b-4234-8000-253f692b204a" (UID: "50a618ec-6a8b-4234-8000-253f692b204a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.851827 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n4tx\" (UniqueName: \"kubernetes.io/projected/50a618ec-6a8b-4234-8000-253f692b204a-kube-api-access-2n4tx\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.851888 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:50 crc kubenswrapper[4970]: I1124 14:13:50.851898 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50a618ec-6a8b-4234-8000-253f692b204a-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.193840 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" event={"ID":"50a618ec-6a8b-4234-8000-253f692b204a","Type":"ContainerDied","Data":"a7b44b0e5b67e59bc863ec27006f0f1374bde4237e951a92d51f267537375677"} Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.193881 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b44b0e5b67e59bc863ec27006f0f1374bde4237e951a92d51f267537375677" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.193903 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fctzs" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.327466 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z"] Nov 24 14:13:51 crc kubenswrapper[4970]: E1124 14:13:51.329350 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50a618ec-6a8b-4234-8000-253f692b204a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329370 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="50a618ec-6a8b-4234-8000-253f692b204a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:51 crc kubenswrapper[4970]: E1124 14:13:51.329401 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="registry-server" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329407 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="registry-server" Nov 24 14:13:51 crc kubenswrapper[4970]: E1124 14:13:51.329422 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="extract-content" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329427 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="extract-content" Nov 24 14:13:51 crc kubenswrapper[4970]: E1124 14:13:51.329438 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="extract-utilities" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329444 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="extract-utilities" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329617 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="50a618ec-6a8b-4234-8000-253f692b204a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.329637 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff82462-7b38-4888-af1f-8d5eb8629c58" containerName="registry-server" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.330347 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.333123 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.333379 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.333482 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.333529 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.345956 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z"] Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.468567 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhfbl\" (UniqueName: \"kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.468670 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.468816 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.570386 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.570496 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.570586 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhfbl\" (UniqueName: \"kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.577044 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.578150 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.592760 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhfbl\" (UniqueName: \"kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:51 crc kubenswrapper[4970]: I1124 14:13:51.654102 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:13:52 crc kubenswrapper[4970]: I1124 14:13:52.184668 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z"] Nov 24 14:13:52 crc kubenswrapper[4970]: W1124 14:13:52.188078 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35b1536c_4ac4_42f5_9f1f_09838c75d1b7.slice/crio-aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38 WatchSource:0}: Error finding container aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38: Status 404 returned error can't find the container with id aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38 Nov 24 14:13:52 crc kubenswrapper[4970]: I1124 14:13:52.203470 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" event={"ID":"35b1536c-4ac4-42f5-9f1f-09838c75d1b7","Type":"ContainerStarted","Data":"aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38"} Nov 24 14:13:53 crc kubenswrapper[4970]: I1124 14:13:53.212924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" event={"ID":"35b1536c-4ac4-42f5-9f1f-09838c75d1b7","Type":"ContainerStarted","Data":"33ed862e0e0bee41560305e58f8961a04ad563bbfe70363ad4faeb4de3413ee0"} Nov 24 14:13:53 crc kubenswrapper[4970]: I1124 14:13:53.237175 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" podStartSLOduration=1.475858985 podStartE2EDuration="2.237152785s" podCreationTimestamp="2025-11-24 14:13:51 +0000 UTC" firstStartedPulling="2025-11-24 14:13:52.190894132 +0000 UTC m=+1647.478651435" lastFinishedPulling="2025-11-24 14:13:52.952187932 +0000 UTC m=+1648.239945235" observedRunningTime="2025-11-24 14:13:53.227351829 +0000 UTC m=+1648.515109142" watchObservedRunningTime="2025-11-24 14:13:53.237152785 +0000 UTC m=+1648.524910088" Nov 24 14:13:58 crc kubenswrapper[4970]: I1124 14:13:58.470631 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:13:58 crc kubenswrapper[4970]: E1124 14:13:58.471743 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:14:11 crc kubenswrapper[4970]: I1124 14:14:11.470466 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:14:11 crc kubenswrapper[4970]: E1124 14:14:11.471149 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:14:17 crc kubenswrapper[4970]: I1124 14:14:17.050697 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7mjd6"] Nov 24 14:14:17 crc kubenswrapper[4970]: I1124 14:14:17.059964 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7mjd6"] Nov 24 14:14:17 crc kubenswrapper[4970]: I1124 14:14:17.482526 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a45bf61d-f620-4455-b4b0-6d0973225fdb" path="/var/lib/kubelet/pods/a45bf61d-f620-4455-b4b0-6d0973225fdb/volumes" Nov 24 14:14:24 crc kubenswrapper[4970]: I1124 14:14:24.470903 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:14:24 crc kubenswrapper[4970]: E1124 14:14:24.472016 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.055083 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bqxgx"] Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.064868 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjs98"] Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.073978 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-gjs98"] Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.082645 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bqxgx"] Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.469952 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:14:39 crc kubenswrapper[4970]: E1124 14:14:39.470189 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.483069 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="478cab1a-9f02-4f21-bae0-5f11ddc62665" path="/var/lib/kubelet/pods/478cab1a-9f02-4f21-bae0-5f11ddc62665/volumes" Nov 24 14:14:39 crc kubenswrapper[4970]: I1124 14:14:39.491212 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c96d3bcf-7d2f-4946-a998-2567e91b275c" path="/var/lib/kubelet/pods/c96d3bcf-7d2f-4946-a998-2567e91b275c/volumes" Nov 24 14:14:41 crc kubenswrapper[4970]: I1124 14:14:41.693835 4970 generic.go:334] "Generic (PLEG): container finished" podID="35b1536c-4ac4-42f5-9f1f-09838c75d1b7" containerID="33ed862e0e0bee41560305e58f8961a04ad563bbfe70363ad4faeb4de3413ee0" exitCode=0 Nov 24 14:14:41 crc kubenswrapper[4970]: I1124 14:14:41.693944 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" event={"ID":"35b1536c-4ac4-42f5-9f1f-09838c75d1b7","Type":"ContainerDied","Data":"33ed862e0e0bee41560305e58f8961a04ad563bbfe70363ad4faeb4de3413ee0"} Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.118126 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.220820 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory\") pod \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.220936 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key\") pod \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.221077 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhfbl\" (UniqueName: \"kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl\") pod \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\" (UID: \"35b1536c-4ac4-42f5-9f1f-09838c75d1b7\") " Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.226978 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl" (OuterVolumeSpecName: "kube-api-access-lhfbl") pod "35b1536c-4ac4-42f5-9f1f-09838c75d1b7" (UID: "35b1536c-4ac4-42f5-9f1f-09838c75d1b7"). InnerVolumeSpecName "kube-api-access-lhfbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.252753 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35b1536c-4ac4-42f5-9f1f-09838c75d1b7" (UID: "35b1536c-4ac4-42f5-9f1f-09838c75d1b7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.253836 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory" (OuterVolumeSpecName: "inventory") pod "35b1536c-4ac4-42f5-9f1f-09838c75d1b7" (UID: "35b1536c-4ac4-42f5-9f1f-09838c75d1b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.322920 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhfbl\" (UniqueName: \"kubernetes.io/projected/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-kube-api-access-lhfbl\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.322968 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.322983 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35b1536c-4ac4-42f5-9f1f-09838c75d1b7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.714143 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" event={"ID":"35b1536c-4ac4-42f5-9f1f-09838c75d1b7","Type":"ContainerDied","Data":"aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38"} Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.714185 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa467b8ad10d70110dc2d678ef31f61ac6fade3b4969fb0c6f6d3b09acbe2c38" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.714283 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.815298 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pxxvx"] Nov 24 14:14:43 crc kubenswrapper[4970]: E1124 14:14:43.815966 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35b1536c-4ac4-42f5-9f1f-09838c75d1b7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.815996 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="35b1536c-4ac4-42f5-9f1f-09838c75d1b7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.816321 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="35b1536c-4ac4-42f5-9f1f-09838c75d1b7" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.817316 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.819472 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.820165 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.821030 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.821103 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.828195 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pxxvx"] Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.934165 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.934422 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg6xm\" (UniqueName: \"kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:43 crc kubenswrapper[4970]: I1124 14:14:43.934545 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.036351 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.036445 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.036523 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg6xm\" (UniqueName: \"kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.042377 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.043100 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.072654 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg6xm\" (UniqueName: \"kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm\") pod \"ssh-known-hosts-edpm-deployment-pxxvx\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.137672 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.443640 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-pxxvx"] Nov 24 14:14:44 crc kubenswrapper[4970]: I1124 14:14:44.725388 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" event={"ID":"6ac1b7b7-0e52-40cc-bd5e-042430621073","Type":"ContainerStarted","Data":"a7d2d31595dfa852112e77b98b06049c3d4795cfe8c6c9dbdf0ab984dbe84306"} Nov 24 14:14:45 crc kubenswrapper[4970]: I1124 14:14:45.734451 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" event={"ID":"6ac1b7b7-0e52-40cc-bd5e-042430621073","Type":"ContainerStarted","Data":"43778bb71dd75bae3f71a859d851e8f20e8065d1fa0048a94b540671ab353c48"} Nov 24 14:14:45 crc kubenswrapper[4970]: I1124 14:14:45.760199 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" podStartSLOduration=2.082922694 podStartE2EDuration="2.76017393s" podCreationTimestamp="2025-11-24 14:14:43 +0000 UTC" firstStartedPulling="2025-11-24 14:14:44.452132837 +0000 UTC m=+1699.739890130" lastFinishedPulling="2025-11-24 14:14:45.129384073 +0000 UTC m=+1700.417141366" observedRunningTime="2025-11-24 14:14:45.755293713 +0000 UTC m=+1701.043051066" watchObservedRunningTime="2025-11-24 14:14:45.76017393 +0000 UTC m=+1701.047931303" Nov 24 14:14:46 crc kubenswrapper[4970]: I1124 14:14:46.509658 4970 scope.go:117] "RemoveContainer" containerID="7c43ed5de706ff312a95ad504212c825b98a056620129c5b1ae8b01e5d1d16b9" Nov 24 14:14:46 crc kubenswrapper[4970]: I1124 14:14:46.552465 4970 scope.go:117] "RemoveContainer" containerID="040259198df7c066f25b1d0a6a41be8acab97047677923e8128574224d415942" Nov 24 14:14:46 crc kubenswrapper[4970]: I1124 14:14:46.597043 4970 scope.go:117] "RemoveContainer" containerID="11497a925aa7d43af4d50072df811a800398b5da23e10542aa3b5c99f3652825" Nov 24 14:14:51 crc kubenswrapper[4970]: I1124 14:14:51.470642 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:14:51 crc kubenswrapper[4970]: E1124 14:14:51.471618 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:14:52 crc kubenswrapper[4970]: I1124 14:14:52.797121 4970 generic.go:334] "Generic (PLEG): container finished" podID="6ac1b7b7-0e52-40cc-bd5e-042430621073" containerID="43778bb71dd75bae3f71a859d851e8f20e8065d1fa0048a94b540671ab353c48" exitCode=0 Nov 24 14:14:52 crc kubenswrapper[4970]: I1124 14:14:52.797221 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" event={"ID":"6ac1b7b7-0e52-40cc-bd5e-042430621073","Type":"ContainerDied","Data":"43778bb71dd75bae3f71a859d851e8f20e8065d1fa0048a94b540671ab353c48"} Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.217173 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.327274 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") pod \"6ac1b7b7-0e52-40cc-bd5e-042430621073\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.327525 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0\") pod \"6ac1b7b7-0e52-40cc-bd5e-042430621073\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.327630 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg6xm\" (UniqueName: \"kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm\") pod \"6ac1b7b7-0e52-40cc-bd5e-042430621073\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.335613 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm" (OuterVolumeSpecName: "kube-api-access-cg6xm") pod "6ac1b7b7-0e52-40cc-bd5e-042430621073" (UID: "6ac1b7b7-0e52-40cc-bd5e-042430621073"). InnerVolumeSpecName "kube-api-access-cg6xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:14:54 crc kubenswrapper[4970]: E1124 14:14:54.358936 4970 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam podName:6ac1b7b7-0e52-40cc-bd5e-042430621073 nodeName:}" failed. No retries permitted until 2025-11-24 14:14:54.858906136 +0000 UTC m=+1710.146663429 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ssh-key-openstack-edpm-ipam" (UniqueName: "kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam") pod "6ac1b7b7-0e52-40cc-bd5e-042430621073" (UID: "6ac1b7b7-0e52-40cc-bd5e-042430621073") : error deleting /var/lib/kubelet/pods/6ac1b7b7-0e52-40cc-bd5e-042430621073/volume-subpaths: remove /var/lib/kubelet/pods/6ac1b7b7-0e52-40cc-bd5e-042430621073/volume-subpaths: no such file or directory Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.364144 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6ac1b7b7-0e52-40cc-bd5e-042430621073" (UID: "6ac1b7b7-0e52-40cc-bd5e-042430621073"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.430557 4970 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.430847 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg6xm\" (UniqueName: \"kubernetes.io/projected/6ac1b7b7-0e52-40cc-bd5e-042430621073-kube-api-access-cg6xm\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.821286 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" event={"ID":"6ac1b7b7-0e52-40cc-bd5e-042430621073","Type":"ContainerDied","Data":"a7d2d31595dfa852112e77b98b06049c3d4795cfe8c6c9dbdf0ab984dbe84306"} Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.821656 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7d2d31595dfa852112e77b98b06049c3d4795cfe8c6c9dbdf0ab984dbe84306" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.821519 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-pxxvx" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.893297 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x"] Nov 24 14:14:54 crc kubenswrapper[4970]: E1124 14:14:54.893813 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac1b7b7-0e52-40cc-bd5e-042430621073" containerName="ssh-known-hosts-edpm-deployment" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.893840 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac1b7b7-0e52-40cc-bd5e-042430621073" containerName="ssh-known-hosts-edpm-deployment" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.894082 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac1b7b7-0e52-40cc-bd5e-042430621073" containerName="ssh-known-hosts-edpm-deployment" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.894932 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.906268 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x"] Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.941318 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") pod \"6ac1b7b7-0e52-40cc-bd5e-042430621073\" (UID: \"6ac1b7b7-0e52-40cc-bd5e-042430621073\") " Nov 24 14:14:54 crc kubenswrapper[4970]: I1124 14:14:54.944954 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6ac1b7b7-0e52-40cc-bd5e-042430621073" (UID: "6ac1b7b7-0e52-40cc-bd5e-042430621073"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.043627 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tckzg\" (UniqueName: \"kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.043689 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.043823 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.043895 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6ac1b7b7-0e52-40cc-bd5e-042430621073-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.145428 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.145743 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tckzg\" (UniqueName: \"kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.145853 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.150850 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.151232 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.175676 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tckzg\" (UniqueName: \"kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zb64x\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.222978 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.742529 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x"] Nov 24 14:14:55 crc kubenswrapper[4970]: I1124 14:14:55.838325 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" event={"ID":"03e03094-db8d-4000-84ea-23ace55b1768","Type":"ContainerStarted","Data":"f9fa7a977f4e0414d32b8ff93c04d8cd775db6982508c346ebff359e8ebd3017"} Nov 24 14:14:56 crc kubenswrapper[4970]: I1124 14:14:56.849201 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" event={"ID":"03e03094-db8d-4000-84ea-23ace55b1768","Type":"ContainerStarted","Data":"841f937d63c262b1b6f1f8fe755146a060c11d6daadfae520b3793acecf9d69a"} Nov 24 14:14:56 crc kubenswrapper[4970]: I1124 14:14:56.867854 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" podStartSLOduration=2.387196495 podStartE2EDuration="2.867837876s" podCreationTimestamp="2025-11-24 14:14:54 +0000 UTC" firstStartedPulling="2025-11-24 14:14:55.750936613 +0000 UTC m=+1711.038693906" lastFinishedPulling="2025-11-24 14:14:56.231578004 +0000 UTC m=+1711.519335287" observedRunningTime="2025-11-24 14:14:56.864159052 +0000 UTC m=+1712.151916355" watchObservedRunningTime="2025-11-24 14:14:56.867837876 +0000 UTC m=+1712.155595169" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.131918 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz"] Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.133457 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.137257 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.137328 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.144499 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz"] Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.253445 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.253745 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.253871 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55bhm\" (UniqueName: \"kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.355260 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55bhm\" (UniqueName: \"kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.355418 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.355442 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.356520 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.365745 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.370197 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55bhm\" (UniqueName: \"kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm\") pod \"collect-profiles-29399895-79zqz\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.453172 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:00 crc kubenswrapper[4970]: I1124 14:15:00.925158 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz"] Nov 24 14:15:00 crc kubenswrapper[4970]: W1124 14:15:00.929297 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0305d3f_93ef_4f25_8784_bc01e7209861.slice/crio-37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae WatchSource:0}: Error finding container 37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae: Status 404 returned error can't find the container with id 37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae Nov 24 14:15:01 crc kubenswrapper[4970]: I1124 14:15:01.905457 4970 generic.go:334] "Generic (PLEG): container finished" podID="c0305d3f-93ef-4f25-8784-bc01e7209861" containerID="161292dcfdf28ad1cca09988dca5812a7e02c5644f067be27e5690a824cf63ef" exitCode=0 Nov 24 14:15:01 crc kubenswrapper[4970]: I1124 14:15:01.905508 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" event={"ID":"c0305d3f-93ef-4f25-8784-bc01e7209861","Type":"ContainerDied","Data":"161292dcfdf28ad1cca09988dca5812a7e02c5644f067be27e5690a824cf63ef"} Nov 24 14:15:01 crc kubenswrapper[4970]: I1124 14:15:01.905803 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" event={"ID":"c0305d3f-93ef-4f25-8784-bc01e7209861","Type":"ContainerStarted","Data":"37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae"} Nov 24 14:15:02 crc kubenswrapper[4970]: I1124 14:15:02.470099 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:15:02 crc kubenswrapper[4970]: E1124 14:15:02.470648 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.272830 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.412618 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55bhm\" (UniqueName: \"kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm\") pod \"c0305d3f-93ef-4f25-8784-bc01e7209861\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.412964 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume\") pod \"c0305d3f-93ef-4f25-8784-bc01e7209861\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.413115 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume\") pod \"c0305d3f-93ef-4f25-8784-bc01e7209861\" (UID: \"c0305d3f-93ef-4f25-8784-bc01e7209861\") " Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.413561 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume" (OuterVolumeSpecName: "config-volume") pod "c0305d3f-93ef-4f25-8784-bc01e7209861" (UID: "c0305d3f-93ef-4f25-8784-bc01e7209861"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.418180 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c0305d3f-93ef-4f25-8784-bc01e7209861" (UID: "c0305d3f-93ef-4f25-8784-bc01e7209861"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.419063 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm" (OuterVolumeSpecName: "kube-api-access-55bhm") pod "c0305d3f-93ef-4f25-8784-bc01e7209861" (UID: "c0305d3f-93ef-4f25-8784-bc01e7209861"). InnerVolumeSpecName "kube-api-access-55bhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.515550 4970 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0305d3f-93ef-4f25-8784-bc01e7209861-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.515605 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55bhm\" (UniqueName: \"kubernetes.io/projected/c0305d3f-93ef-4f25-8784-bc01e7209861-kube-api-access-55bhm\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.515619 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0305d3f-93ef-4f25-8784-bc01e7209861-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.935412 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" event={"ID":"c0305d3f-93ef-4f25-8784-bc01e7209861","Type":"ContainerDied","Data":"37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae"} Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.935461 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37fb6e73c52d44ae619c6b17582c73c10214cf7cc299aa4d0ddceec5a5c289ae" Nov 24 14:15:03 crc kubenswrapper[4970]: I1124 14:15:03.935543 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399895-79zqz" Nov 24 14:15:04 crc kubenswrapper[4970]: I1124 14:15:04.947892 4970 generic.go:334] "Generic (PLEG): container finished" podID="03e03094-db8d-4000-84ea-23ace55b1768" containerID="841f937d63c262b1b6f1f8fe755146a060c11d6daadfae520b3793acecf9d69a" exitCode=0 Nov 24 14:15:04 crc kubenswrapper[4970]: I1124 14:15:04.947978 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" event={"ID":"03e03094-db8d-4000-84ea-23ace55b1768","Type":"ContainerDied","Data":"841f937d63c262b1b6f1f8fe755146a060c11d6daadfae520b3793acecf9d69a"} Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.336879 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.468677 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tckzg\" (UniqueName: \"kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg\") pod \"03e03094-db8d-4000-84ea-23ace55b1768\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.470155 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key\") pod \"03e03094-db8d-4000-84ea-23ace55b1768\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.470291 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory\") pod \"03e03094-db8d-4000-84ea-23ace55b1768\" (UID: \"03e03094-db8d-4000-84ea-23ace55b1768\") " Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.474932 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg" (OuterVolumeSpecName: "kube-api-access-tckzg") pod "03e03094-db8d-4000-84ea-23ace55b1768" (UID: "03e03094-db8d-4000-84ea-23ace55b1768"). InnerVolumeSpecName "kube-api-access-tckzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.500029 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory" (OuterVolumeSpecName: "inventory") pod "03e03094-db8d-4000-84ea-23ace55b1768" (UID: "03e03094-db8d-4000-84ea-23ace55b1768"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.501523 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "03e03094-db8d-4000-84ea-23ace55b1768" (UID: "03e03094-db8d-4000-84ea-23ace55b1768"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.572198 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.572238 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03e03094-db8d-4000-84ea-23ace55b1768-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:06 crc kubenswrapper[4970]: I1124 14:15:06.572251 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tckzg\" (UniqueName: \"kubernetes.io/projected/03e03094-db8d-4000-84ea-23ace55b1768-kube-api-access-tckzg\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.041767 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" event={"ID":"03e03094-db8d-4000-84ea-23ace55b1768","Type":"ContainerDied","Data":"f9fa7a977f4e0414d32b8ff93c04d8cd775db6982508c346ebff359e8ebd3017"} Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.042085 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9fa7a977f4e0414d32b8ff93c04d8cd775db6982508c346ebff359e8ebd3017" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.041861 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zb64x" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.088725 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m"] Nov 24 14:15:07 crc kubenswrapper[4970]: E1124 14:15:07.089077 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0305d3f-93ef-4f25-8784-bc01e7209861" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.089092 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0305d3f-93ef-4f25-8784-bc01e7209861" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4970]: E1124 14:15:07.089118 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03e03094-db8d-4000-84ea-23ace55b1768" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.089126 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="03e03094-db8d-4000-84ea-23ace55b1768" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.089301 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="03e03094-db8d-4000-84ea-23ace55b1768" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.089320 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0305d3f-93ef-4f25-8784-bc01e7209861" containerName="collect-profiles" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.089895 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.092595 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.093162 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.109364 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.113379 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.132104 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m"] Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.235567 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.235742 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf9cw\" (UniqueName: \"kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.235809 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.337678 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf9cw\" (UniqueName: \"kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.337751 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.337817 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.345716 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.347437 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.358659 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf9cw\" (UniqueName: \"kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.411568 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:07 crc kubenswrapper[4970]: I1124 14:15:07.976906 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m"] Nov 24 14:15:08 crc kubenswrapper[4970]: I1124 14:15:08.050432 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" event={"ID":"ee264c1e-7f6a-4dc1-b406-aced8f917394","Type":"ContainerStarted","Data":"4d2c82061d5a78efcb35c8d61d10465c82974c0e58ac79d7241a5341a2325b1c"} Nov 24 14:15:09 crc kubenswrapper[4970]: I1124 14:15:09.070769 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" event={"ID":"ee264c1e-7f6a-4dc1-b406-aced8f917394","Type":"ContainerStarted","Data":"53a1755af07ed06d8ab78311cdcf3cf20a689b325edd3add78e9cd69aab6e5cb"} Nov 24 14:15:09 crc kubenswrapper[4970]: I1124 14:15:09.097015 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" podStartSLOduration=1.590213296 podStartE2EDuration="2.096995773s" podCreationTimestamp="2025-11-24 14:15:07 +0000 UTC" firstStartedPulling="2025-11-24 14:15:07.982815577 +0000 UTC m=+1723.270572860" lastFinishedPulling="2025-11-24 14:15:08.489598034 +0000 UTC m=+1723.777355337" observedRunningTime="2025-11-24 14:15:09.091515869 +0000 UTC m=+1724.379273162" watchObservedRunningTime="2025-11-24 14:15:09.096995773 +0000 UTC m=+1724.384753066" Nov 24 14:15:16 crc kubenswrapper[4970]: I1124 14:15:16.469941 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:15:16 crc kubenswrapper[4970]: E1124 14:15:16.470763 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:15:19 crc kubenswrapper[4970]: I1124 14:15:19.183658 4970 generic.go:334] "Generic (PLEG): container finished" podID="ee264c1e-7f6a-4dc1-b406-aced8f917394" containerID="53a1755af07ed06d8ab78311cdcf3cf20a689b325edd3add78e9cd69aab6e5cb" exitCode=0 Nov 24 14:15:19 crc kubenswrapper[4970]: I1124 14:15:19.183726 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" event={"ID":"ee264c1e-7f6a-4dc1-b406-aced8f917394","Type":"ContainerDied","Data":"53a1755af07ed06d8ab78311cdcf3cf20a689b325edd3add78e9cd69aab6e5cb"} Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.606320 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.691371 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key\") pod \"ee264c1e-7f6a-4dc1-b406-aced8f917394\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.691505 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory\") pod \"ee264c1e-7f6a-4dc1-b406-aced8f917394\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.691559 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf9cw\" (UniqueName: \"kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw\") pod \"ee264c1e-7f6a-4dc1-b406-aced8f917394\" (UID: \"ee264c1e-7f6a-4dc1-b406-aced8f917394\") " Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.696539 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw" (OuterVolumeSpecName: "kube-api-access-vf9cw") pod "ee264c1e-7f6a-4dc1-b406-aced8f917394" (UID: "ee264c1e-7f6a-4dc1-b406-aced8f917394"). InnerVolumeSpecName "kube-api-access-vf9cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.723771 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory" (OuterVolumeSpecName: "inventory") pod "ee264c1e-7f6a-4dc1-b406-aced8f917394" (UID: "ee264c1e-7f6a-4dc1-b406-aced8f917394"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.725021 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ee264c1e-7f6a-4dc1-b406-aced8f917394" (UID: "ee264c1e-7f6a-4dc1-b406-aced8f917394"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.793425 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.793463 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf9cw\" (UniqueName: \"kubernetes.io/projected/ee264c1e-7f6a-4dc1-b406-aced8f917394-kube-api-access-vf9cw\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:20 crc kubenswrapper[4970]: I1124 14:15:20.793474 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ee264c1e-7f6a-4dc1-b406-aced8f917394-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.207780 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" event={"ID":"ee264c1e-7f6a-4dc1-b406-aced8f917394","Type":"ContainerDied","Data":"4d2c82061d5a78efcb35c8d61d10465c82974c0e58ac79d7241a5341a2325b1c"} Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.207839 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d2c82061d5a78efcb35c8d61d10465c82974c0e58ac79d7241a5341a2325b1c" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.207864 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.304550 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p"] Nov 24 14:15:21 crc kubenswrapper[4970]: E1124 14:15:21.304923 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee264c1e-7f6a-4dc1-b406-aced8f917394" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.304940 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee264c1e-7f6a-4dc1-b406-aced8f917394" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.305130 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee264c1e-7f6a-4dc1-b406-aced8f917394" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.305774 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.307740 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.311659 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.312057 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.312699 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.312982 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.313235 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.313675 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.314215 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.332760 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p"] Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.403390 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.403892 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.403973 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.403997 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404030 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404247 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404330 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404539 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404689 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404824 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404867 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl676\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404905 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.404939 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.405073 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507430 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507507 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl676\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507558 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507603 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507656 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507724 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507781 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507816 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507842 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507871 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507917 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.507952 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.508018 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.508060 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.513801 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.513821 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.514510 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.514807 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.515443 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.516177 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.516488 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.516894 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.517855 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.518200 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.518799 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.519598 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.520543 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.531809 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl676\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:21 crc kubenswrapper[4970]: I1124 14:15:21.631637 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:15:22 crc kubenswrapper[4970]: W1124 14:15:22.190773 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda87a7cba_b0be_443b_8393_6f0f21e1ccb9.slice/crio-72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e WatchSource:0}: Error finding container 72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e: Status 404 returned error can't find the container with id 72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e Nov 24 14:15:22 crc kubenswrapper[4970]: I1124 14:15:22.191742 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p"] Nov 24 14:15:22 crc kubenswrapper[4970]: I1124 14:15:22.219803 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" event={"ID":"a87a7cba-b0be-443b-8393-6f0f21e1ccb9","Type":"ContainerStarted","Data":"72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e"} Nov 24 14:15:23 crc kubenswrapper[4970]: I1124 14:15:23.229008 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" event={"ID":"a87a7cba-b0be-443b-8393-6f0f21e1ccb9","Type":"ContainerStarted","Data":"aa24527179a511eae1e55afffaf6232ee26ba407d098112ce84edac8cf02d91e"} Nov 24 14:15:23 crc kubenswrapper[4970]: I1124 14:15:23.255690 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" podStartSLOduration=1.759618605 podStartE2EDuration="2.255670559s" podCreationTimestamp="2025-11-24 14:15:21 +0000 UTC" firstStartedPulling="2025-11-24 14:15:22.193802587 +0000 UTC m=+1737.481559880" lastFinishedPulling="2025-11-24 14:15:22.689854491 +0000 UTC m=+1737.977611834" observedRunningTime="2025-11-24 14:15:23.249901077 +0000 UTC m=+1738.537658380" watchObservedRunningTime="2025-11-24 14:15:23.255670559 +0000 UTC m=+1738.543427852" Nov 24 14:15:25 crc kubenswrapper[4970]: I1124 14:15:25.038707 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-chg79"] Nov 24 14:15:25 crc kubenswrapper[4970]: I1124 14:15:25.048646 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-chg79"] Nov 24 14:15:25 crc kubenswrapper[4970]: I1124 14:15:25.489684 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e87b7aad-6180-461c-981c-f1a1c8be8beb" path="/var/lib/kubelet/pods/e87b7aad-6180-461c-981c-f1a1c8be8beb/volumes" Nov 24 14:15:29 crc kubenswrapper[4970]: I1124 14:15:29.470716 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:15:29 crc kubenswrapper[4970]: E1124 14:15:29.471471 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:15:40 crc kubenswrapper[4970]: I1124 14:15:40.471703 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:15:40 crc kubenswrapper[4970]: E1124 14:15:40.472812 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:15:46 crc kubenswrapper[4970]: I1124 14:15:46.703626 4970 scope.go:117] "RemoveContainer" containerID="853d1adce379e23a214bf4f8822859a9feae49e1dca8e86f02431eb9ee9b0c2e" Nov 24 14:15:55 crc kubenswrapper[4970]: I1124 14:15:55.475420 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:15:55 crc kubenswrapper[4970]: E1124 14:15:55.476262 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:16:01 crc kubenswrapper[4970]: I1124 14:16:01.594077 4970 generic.go:334] "Generic (PLEG): container finished" podID="a87a7cba-b0be-443b-8393-6f0f21e1ccb9" containerID="aa24527179a511eae1e55afffaf6232ee26ba407d098112ce84edac8cf02d91e" exitCode=0 Nov 24 14:16:01 crc kubenswrapper[4970]: I1124 14:16:01.594160 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" event={"ID":"a87a7cba-b0be-443b-8393-6f0f21e1ccb9","Type":"ContainerDied","Data":"aa24527179a511eae1e55afffaf6232ee26ba407d098112ce84edac8cf02d91e"} Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.093853 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207180 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207430 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207469 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207496 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207521 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207544 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207559 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207623 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207650 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207701 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207764 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207806 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl676\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207830 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.207854 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle\") pod \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\" (UID: \"a87a7cba-b0be-443b-8393-6f0f21e1ccb9\") " Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.212888 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.213642 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.214135 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.214222 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.214277 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.215739 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.216043 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.216570 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.225814 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676" (OuterVolumeSpecName: "kube-api-access-kl676") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "kube-api-access-kl676". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.226983 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.228361 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.230619 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.239267 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory" (OuterVolumeSpecName: "inventory") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.242713 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a87a7cba-b0be-443b-8393-6f0f21e1ccb9" (UID: "a87a7cba-b0be-443b-8393-6f0f21e1ccb9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.309901 4970 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310132 4970 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310149 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310164 4970 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310175 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310188 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310198 4970 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310209 4970 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310220 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310234 4970 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310244 4970 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310254 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310264 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.310274 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl676\" (UniqueName: \"kubernetes.io/projected/a87a7cba-b0be-443b-8393-6f0f21e1ccb9-kube-api-access-kl676\") on node \"crc\" DevicePath \"\"" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.614256 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" event={"ID":"a87a7cba-b0be-443b-8393-6f0f21e1ccb9","Type":"ContainerDied","Data":"72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e"} Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.614301 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.614306 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72c78b2e44d5353a0ffd5acd4c2cc2299b4e99e9652e22ccfd6414e9152dd59e" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.727298 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t"] Nov 24 14:16:03 crc kubenswrapper[4970]: E1124 14:16:03.727760 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87a7cba-b0be-443b-8393-6f0f21e1ccb9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.727781 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87a7cba-b0be-443b-8393-6f0f21e1ccb9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.727972 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87a7cba-b0be-443b-8393-6f0f21e1ccb9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.728626 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.731315 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.731443 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.731486 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.732868 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.732869 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.738008 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t"] Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.818180 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.818473 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.818536 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.818642 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpsjw\" (UniqueName: \"kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.818809 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.920445 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.920490 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.920528 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpsjw\" (UniqueName: \"kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.920601 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.920693 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.921880 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.925884 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.927115 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.927893 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:03 crc kubenswrapper[4970]: I1124 14:16:03.945431 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpsjw\" (UniqueName: \"kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7dr7t\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:04 crc kubenswrapper[4970]: I1124 14:16:04.043128 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:16:04 crc kubenswrapper[4970]: I1124 14:16:04.591412 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t"] Nov 24 14:16:04 crc kubenswrapper[4970]: I1124 14:16:04.623914 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" event={"ID":"1843fb9f-02d8-4021-8ef5-f5416f439ef5","Type":"ContainerStarted","Data":"917cb13cd61e5d89908141e582e09829d48c2c2b3ae06876bd339eb1e0f5acc1"} Nov 24 14:16:05 crc kubenswrapper[4970]: I1124 14:16:05.632883 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" event={"ID":"1843fb9f-02d8-4021-8ef5-f5416f439ef5","Type":"ContainerStarted","Data":"b1d67fe4cb952fbe376e650b57e6d9e80e50c62f75e151f59cb398e44f675c69"} Nov 24 14:16:05 crc kubenswrapper[4970]: I1124 14:16:05.660151 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" podStartSLOduration=2.07443994 podStartE2EDuration="2.660129384s" podCreationTimestamp="2025-11-24 14:16:03 +0000 UTC" firstStartedPulling="2025-11-24 14:16:04.598193418 +0000 UTC m=+1779.885950711" lastFinishedPulling="2025-11-24 14:16:05.183882842 +0000 UTC m=+1780.471640155" observedRunningTime="2025-11-24 14:16:05.650668517 +0000 UTC m=+1780.938425820" watchObservedRunningTime="2025-11-24 14:16:05.660129384 +0000 UTC m=+1780.947886677" Nov 24 14:16:10 crc kubenswrapper[4970]: I1124 14:16:10.471027 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:16:10 crc kubenswrapper[4970]: E1124 14:16:10.472130 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:16:24 crc kubenswrapper[4970]: I1124 14:16:24.470729 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:16:24 crc kubenswrapper[4970]: I1124 14:16:24.828462 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073"} Nov 24 14:16:46 crc kubenswrapper[4970]: I1124 14:16:46.814954 4970 scope.go:117] "RemoveContainer" containerID="2454be05bb3a43170c82a837d224171f87eff0e64effccc7f1cfd99fb321dd04" Nov 24 14:16:46 crc kubenswrapper[4970]: I1124 14:16:46.849933 4970 scope.go:117] "RemoveContainer" containerID="61618aec90053c020a383f53d62e916e97dd79ce0658925b77410a2a756ae419" Nov 24 14:16:46 crc kubenswrapper[4970]: I1124 14:16:46.885720 4970 scope.go:117] "RemoveContainer" containerID="06e530c66f85f0747bbfb32721e9b5a87992b83388d36f67926573fba693b0ae" Nov 24 14:17:06 crc kubenswrapper[4970]: I1124 14:17:06.260962 4970 generic.go:334] "Generic (PLEG): container finished" podID="1843fb9f-02d8-4021-8ef5-f5416f439ef5" containerID="b1d67fe4cb952fbe376e650b57e6d9e80e50c62f75e151f59cb398e44f675c69" exitCode=0 Nov 24 14:17:06 crc kubenswrapper[4970]: I1124 14:17:06.261069 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" event={"ID":"1843fb9f-02d8-4021-8ef5-f5416f439ef5","Type":"ContainerDied","Data":"b1d67fe4cb952fbe376e650b57e6d9e80e50c62f75e151f59cb398e44f675c69"} Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.685344 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.773544 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpsjw\" (UniqueName: \"kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw\") pod \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.773671 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory\") pod \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.773770 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0\") pod \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.773839 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle\") pod \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.773865 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key\") pod \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\" (UID: \"1843fb9f-02d8-4021-8ef5-f5416f439ef5\") " Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.779840 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw" (OuterVolumeSpecName: "kube-api-access-tpsjw") pod "1843fb9f-02d8-4021-8ef5-f5416f439ef5" (UID: "1843fb9f-02d8-4021-8ef5-f5416f439ef5"). InnerVolumeSpecName "kube-api-access-tpsjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.781678 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1843fb9f-02d8-4021-8ef5-f5416f439ef5" (UID: "1843fb9f-02d8-4021-8ef5-f5416f439ef5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.805165 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1843fb9f-02d8-4021-8ef5-f5416f439ef5" (UID: "1843fb9f-02d8-4021-8ef5-f5416f439ef5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.805428 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1843fb9f-02d8-4021-8ef5-f5416f439ef5" (UID: "1843fb9f-02d8-4021-8ef5-f5416f439ef5"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.813159 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory" (OuterVolumeSpecName: "inventory") pod "1843fb9f-02d8-4021-8ef5-f5416f439ef5" (UID: "1843fb9f-02d8-4021-8ef5-f5416f439ef5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.875863 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.875894 4970 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.875905 4970 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.875912 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1843fb9f-02d8-4021-8ef5-f5416f439ef5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:07 crc kubenswrapper[4970]: I1124 14:17:07.875921 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpsjw\" (UniqueName: \"kubernetes.io/projected/1843fb9f-02d8-4021-8ef5-f5416f439ef5-kube-api-access-tpsjw\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.281553 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" event={"ID":"1843fb9f-02d8-4021-8ef5-f5416f439ef5","Type":"ContainerDied","Data":"917cb13cd61e5d89908141e582e09829d48c2c2b3ae06876bd339eb1e0f5acc1"} Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.281623 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="917cb13cd61e5d89908141e582e09829d48c2c2b3ae06876bd339eb1e0f5acc1" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.281631 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7dr7t" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.379462 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw"] Nov 24 14:17:08 crc kubenswrapper[4970]: E1124 14:17:08.379920 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1843fb9f-02d8-4021-8ef5-f5416f439ef5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.379944 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1843fb9f-02d8-4021-8ef5-f5416f439ef5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.380132 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="1843fb9f-02d8-4021-8ef5-f5416f439ef5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.380781 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.386456 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.386785 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.386903 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.387131 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.387334 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.387419 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.395051 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw"] Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486325 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486414 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486477 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486692 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486899 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b9tp\" (UniqueName: \"kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.486977 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.589080 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.589206 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.590870 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.590989 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.591035 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.591183 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b9tp\" (UniqueName: \"kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.596195 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.599161 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.599514 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.599641 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.600987 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.610518 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b9tp\" (UniqueName: \"kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:08 crc kubenswrapper[4970]: I1124 14:17:08.700618 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:09 crc kubenswrapper[4970]: I1124 14:17:09.256538 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw"] Nov 24 14:17:09 crc kubenswrapper[4970]: I1124 14:17:09.270631 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:17:09 crc kubenswrapper[4970]: I1124 14:17:09.294003 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" event={"ID":"c1a37385-6027-4ee8-ad98-269f93d2c46d","Type":"ContainerStarted","Data":"bbd7f2e0c19c900b49efe86a5f399a8135c299f253ac4b7a06b163a3805dc1b5"} Nov 24 14:17:11 crc kubenswrapper[4970]: I1124 14:17:11.330903 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" event={"ID":"c1a37385-6027-4ee8-ad98-269f93d2c46d","Type":"ContainerStarted","Data":"7eb8dd183be84b4a36f75d6e006f5b3909362975c3d70678f940ddd61ca460c5"} Nov 24 14:17:11 crc kubenswrapper[4970]: I1124 14:17:11.361679 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" podStartSLOduration=2.6101930639999997 podStartE2EDuration="3.361657406s" podCreationTimestamp="2025-11-24 14:17:08 +0000 UTC" firstStartedPulling="2025-11-24 14:17:09.270335132 +0000 UTC m=+1844.558092425" lastFinishedPulling="2025-11-24 14:17:10.021799474 +0000 UTC m=+1845.309556767" observedRunningTime="2025-11-24 14:17:11.354859064 +0000 UTC m=+1846.642616357" watchObservedRunningTime="2025-11-24 14:17:11.361657406 +0000 UTC m=+1846.649414699" Nov 24 14:17:56 crc kubenswrapper[4970]: I1124 14:17:56.045259 4970 generic.go:334] "Generic (PLEG): container finished" podID="c1a37385-6027-4ee8-ad98-269f93d2c46d" containerID="7eb8dd183be84b4a36f75d6e006f5b3909362975c3d70678f940ddd61ca460c5" exitCode=0 Nov 24 14:17:56 crc kubenswrapper[4970]: I1124 14:17:56.045346 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" event={"ID":"c1a37385-6027-4ee8-ad98-269f93d2c46d","Type":"ContainerDied","Data":"7eb8dd183be84b4a36f75d6e006f5b3909362975c3d70678f940ddd61ca460c5"} Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.423756 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.536403 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.536489 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b9tp\" (UniqueName: \"kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.536570 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.536739 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.536903 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.537560 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key\") pod \"c1a37385-6027-4ee8-ad98-269f93d2c46d\" (UID: \"c1a37385-6027-4ee8-ad98-269f93d2c46d\") " Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.543089 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.543836 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp" (OuterVolumeSpecName: "kube-api-access-8b9tp") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "kube-api-access-8b9tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.565930 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.566538 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.568696 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.572033 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory" (OuterVolumeSpecName: "inventory") pod "c1a37385-6027-4ee8-ad98-269f93d2c46d" (UID: "c1a37385-6027-4ee8-ad98-269f93d2c46d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639790 4970 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639825 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639836 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639845 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b9tp\" (UniqueName: \"kubernetes.io/projected/c1a37385-6027-4ee8-ad98-269f93d2c46d-kube-api-access-8b9tp\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639854 4970 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:57 crc kubenswrapper[4970]: I1124 14:17:57.639866 4970 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c1a37385-6027-4ee8-ad98-269f93d2c46d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.065620 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" event={"ID":"c1a37385-6027-4ee8-ad98-269f93d2c46d","Type":"ContainerDied","Data":"bbd7f2e0c19c900b49efe86a5f399a8135c299f253ac4b7a06b163a3805dc1b5"} Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.065950 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbd7f2e0c19c900b49efe86a5f399a8135c299f253ac4b7a06b163a3805dc1b5" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.065729 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.166940 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv"] Nov 24 14:17:58 crc kubenswrapper[4970]: E1124 14:17:58.167417 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a37385-6027-4ee8-ad98-269f93d2c46d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.167441 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a37385-6027-4ee8-ad98-269f93d2c46d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.167729 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a37385-6027-4ee8-ad98-269f93d2c46d" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.168543 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.170406 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.170818 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.171104 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.172115 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.173033 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.182160 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv"] Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.249049 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.249308 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.249367 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.249401 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swvvd\" (UniqueName: \"kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.249653 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.351411 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.351468 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.351498 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swvvd\" (UniqueName: \"kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.351545 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.351611 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.356103 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.356208 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.356470 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.357745 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.368743 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swvvd\" (UniqueName: \"kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:58 crc kubenswrapper[4970]: I1124 14:17:58.495026 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:17:59 crc kubenswrapper[4970]: I1124 14:17:59.021555 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv"] Nov 24 14:17:59 crc kubenswrapper[4970]: I1124 14:17:59.075175 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" event={"ID":"5dcb0227-fbb2-492f-819b-02dd5de1a728","Type":"ContainerStarted","Data":"1ed2a744925e7f94339f4c293345f41bc4bdc850a2519ec9239de593d85f25be"} Nov 24 14:18:00 crc kubenswrapper[4970]: I1124 14:18:00.086604 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" event={"ID":"5dcb0227-fbb2-492f-819b-02dd5de1a728","Type":"ContainerStarted","Data":"75668996cfbded98f2cdaee223ad3ed2359882b3043f753a489f9de4e5416b4d"} Nov 24 14:18:00 crc kubenswrapper[4970]: I1124 14:18:00.103205 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" podStartSLOduration=1.628258513 podStartE2EDuration="2.103182718s" podCreationTimestamp="2025-11-24 14:17:58 +0000 UTC" firstStartedPulling="2025-11-24 14:17:59.032321982 +0000 UTC m=+1894.320079275" lastFinishedPulling="2025-11-24 14:17:59.507246187 +0000 UTC m=+1894.795003480" observedRunningTime="2025-11-24 14:18:00.10071956 +0000 UTC m=+1895.388476853" watchObservedRunningTime="2025-11-24 14:18:00.103182718 +0000 UTC m=+1895.390940041" Nov 24 14:18:41 crc kubenswrapper[4970]: I1124 14:18:41.204739 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:18:41 crc kubenswrapper[4970]: I1124 14:18:41.220685 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:19:11 crc kubenswrapper[4970]: I1124 14:19:11.203866 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:19:11 crc kubenswrapper[4970]: I1124 14:19:11.204418 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:19:41 crc kubenswrapper[4970]: I1124 14:19:41.204787 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:19:41 crc kubenswrapper[4970]: I1124 14:19:41.205321 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:19:41 crc kubenswrapper[4970]: I1124 14:19:41.205367 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:19:41 crc kubenswrapper[4970]: I1124 14:19:41.206228 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:19:41 crc kubenswrapper[4970]: I1124 14:19:41.206307 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073" gracePeriod=600 Nov 24 14:19:42 crc kubenswrapper[4970]: I1124 14:19:42.033014 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073" exitCode=0 Nov 24 14:19:42 crc kubenswrapper[4970]: I1124 14:19:42.033076 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073"} Nov 24 14:19:42 crc kubenswrapper[4970]: I1124 14:19:42.033533 4970 scope.go:117] "RemoveContainer" containerID="66bb36ab6e15f1ea753431d09c7360d5593b51762a5038d542ac0c224e79f270" Nov 24 14:19:43 crc kubenswrapper[4970]: I1124 14:19:43.044088 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35"} Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.446957 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.449182 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.473135 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.595601 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.595850 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.595904 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td89s\" (UniqueName: \"kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.698865 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.698962 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.698998 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td89s\" (UniqueName: \"kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.699718 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.699749 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.719806 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td89s\" (UniqueName: \"kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s\") pod \"certified-operators-92q8l\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:58 crc kubenswrapper[4970]: I1124 14:20:58.822560 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:20:59 crc kubenswrapper[4970]: I1124 14:20:59.297157 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:20:59 crc kubenswrapper[4970]: I1124 14:20:59.748432 4970 generic.go:334] "Generic (PLEG): container finished" podID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerID="1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b" exitCode=0 Nov 24 14:20:59 crc kubenswrapper[4970]: I1124 14:20:59.748478 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerDied","Data":"1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b"} Nov 24 14:20:59 crc kubenswrapper[4970]: I1124 14:20:59.748512 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerStarted","Data":"942275fb75011bbea63f6d487908f42ceeb9c616f257129fe06bd0aa5bf42844"} Nov 24 14:21:00 crc kubenswrapper[4970]: I1124 14:21:00.768093 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerStarted","Data":"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3"} Nov 24 14:21:01 crc kubenswrapper[4970]: I1124 14:21:01.782848 4970 generic.go:334] "Generic (PLEG): container finished" podID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerID="f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3" exitCode=0 Nov 24 14:21:01 crc kubenswrapper[4970]: I1124 14:21:01.782932 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerDied","Data":"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3"} Nov 24 14:21:02 crc kubenswrapper[4970]: I1124 14:21:02.805377 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerStarted","Data":"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186"} Nov 24 14:21:02 crc kubenswrapper[4970]: I1124 14:21:02.828479 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-92q8l" podStartSLOduration=3.014632743 podStartE2EDuration="4.828453836s" podCreationTimestamp="2025-11-24 14:20:58 +0000 UTC" firstStartedPulling="2025-11-24 14:20:59.751285523 +0000 UTC m=+2075.039042816" lastFinishedPulling="2025-11-24 14:21:01.565106606 +0000 UTC m=+2076.852863909" observedRunningTime="2025-11-24 14:21:02.821257189 +0000 UTC m=+2078.109014482" watchObservedRunningTime="2025-11-24 14:21:02.828453836 +0000 UTC m=+2078.116211129" Nov 24 14:21:08 crc kubenswrapper[4970]: I1124 14:21:08.823195 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:08 crc kubenswrapper[4970]: I1124 14:21:08.823818 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:08 crc kubenswrapper[4970]: I1124 14:21:08.933267 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:08 crc kubenswrapper[4970]: I1124 14:21:08.976155 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:09 crc kubenswrapper[4970]: I1124 14:21:09.172876 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:21:10 crc kubenswrapper[4970]: I1124 14:21:10.909968 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-92q8l" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="registry-server" containerID="cri-o://8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186" gracePeriod=2 Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.437103 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.541431 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td89s\" (UniqueName: \"kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s\") pod \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.541622 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities\") pod \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.541713 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content\") pod \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\" (UID: \"8a815d6b-8c4e-4699-976b-d7c51ca75ee6\") " Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.542800 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities" (OuterVolumeSpecName: "utilities") pod "8a815d6b-8c4e-4699-976b-d7c51ca75ee6" (UID: "8a815d6b-8c4e-4699-976b-d7c51ca75ee6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.548906 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s" (OuterVolumeSpecName: "kube-api-access-td89s") pod "8a815d6b-8c4e-4699-976b-d7c51ca75ee6" (UID: "8a815d6b-8c4e-4699-976b-d7c51ca75ee6"). InnerVolumeSpecName "kube-api-access-td89s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.593417 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a815d6b-8c4e-4699-976b-d7c51ca75ee6" (UID: "8a815d6b-8c4e-4699-976b-d7c51ca75ee6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.643957 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.643998 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.644014 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td89s\" (UniqueName: \"kubernetes.io/projected/8a815d6b-8c4e-4699-976b-d7c51ca75ee6-kube-api-access-td89s\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.938510 4970 generic.go:334] "Generic (PLEG): container finished" podID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerID="8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186" exitCode=0 Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.938601 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-92q8l" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.938625 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerDied","Data":"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186"} Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.938912 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-92q8l" event={"ID":"8a815d6b-8c4e-4699-976b-d7c51ca75ee6","Type":"ContainerDied","Data":"942275fb75011bbea63f6d487908f42ceeb9c616f257129fe06bd0aa5bf42844"} Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.938931 4970 scope.go:117] "RemoveContainer" containerID="8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.959727 4970 scope.go:117] "RemoveContainer" containerID="f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3" Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.975284 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.982444 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-92q8l"] Nov 24 14:21:11 crc kubenswrapper[4970]: I1124 14:21:11.999489 4970 scope.go:117] "RemoveContainer" containerID="1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.029941 4970 scope.go:117] "RemoveContainer" containerID="8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186" Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.030368 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186\": container with ID starting with 8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186 not found: ID does not exist" containerID="8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.030428 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186"} err="failed to get container status \"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186\": rpc error: code = NotFound desc = could not find container \"8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186\": container with ID starting with 8b54c728f314fb49a1f5be416c1aa147c69198401c0b9512fddcb9e08cd73186 not found: ID does not exist" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.030461 4970 scope.go:117] "RemoveContainer" containerID="f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3" Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.030964 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3\": container with ID starting with f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3 not found: ID does not exist" containerID="f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.030999 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3"} err="failed to get container status \"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3\": rpc error: code = NotFound desc = could not find container \"f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3\": container with ID starting with f71b9a9828b5fab6ca0c7e26cc604c90f617e9f4f8c7fa59bf4d8bf161a14ae3 not found: ID does not exist" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.031013 4970 scope.go:117] "RemoveContainer" containerID="1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b" Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.031259 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b\": container with ID starting with 1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b not found: ID does not exist" containerID="1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.031290 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b"} err="failed to get container status \"1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b\": rpc error: code = NotFound desc = could not find container \"1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b\": container with ID starting with 1cd08909fd4d211e8524b7ef4accfda285d6b0a85508cfea98d17cc1974b2c3b not found: ID does not exist" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.978409 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ggqrv"] Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.979595 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="registry-server" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.979686 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="registry-server" Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.979761 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="extract-utilities" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.979890 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="extract-utilities" Nov 24 14:21:12 crc kubenswrapper[4970]: E1124 14:21:12.980012 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="extract-content" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.980228 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="extract-content" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.980524 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" containerName="registry-server" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.986315 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:12 crc kubenswrapper[4970]: I1124 14:21:12.994143 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ggqrv"] Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.072413 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-catalog-content\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.072481 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-utilities\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.072514 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4m5c\" (UniqueName: \"kubernetes.io/projected/002977b6-e9fd-4de3-8753-96a4d041c7fd-kube-api-access-x4m5c\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.174343 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4m5c\" (UniqueName: \"kubernetes.io/projected/002977b6-e9fd-4de3-8753-96a4d041c7fd-kube-api-access-x4m5c\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.174671 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-catalog-content\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.174765 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-utilities\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.175216 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-catalog-content\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.175335 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/002977b6-e9fd-4de3-8753-96a4d041c7fd-utilities\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.198043 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4m5c\" (UniqueName: \"kubernetes.io/projected/002977b6-e9fd-4de3-8753-96a4d041c7fd-kube-api-access-x4m5c\") pod \"redhat-operators-ggqrv\" (UID: \"002977b6-e9fd-4de3-8753-96a4d041c7fd\") " pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.311134 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.490972 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a815d6b-8c4e-4699-976b-d7c51ca75ee6" path="/var/lib/kubelet/pods/8a815d6b-8c4e-4699-976b-d7c51ca75ee6/volumes" Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.758338 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ggqrv"] Nov 24 14:21:13 crc kubenswrapper[4970]: I1124 14:21:13.956606 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggqrv" event={"ID":"002977b6-e9fd-4de3-8753-96a4d041c7fd","Type":"ContainerStarted","Data":"5f8fde5b491818d5ede83334718b5794411224d9599eeeff4a75b357fe67aaf9"} Nov 24 14:21:14 crc kubenswrapper[4970]: I1124 14:21:14.972848 4970 generic.go:334] "Generic (PLEG): container finished" podID="002977b6-e9fd-4de3-8753-96a4d041c7fd" containerID="2ad7de06da98f28d976dd679bc176948cb700941bd05de91282d062f6fd0786e" exitCode=0 Nov 24 14:21:14 crc kubenswrapper[4970]: I1124 14:21:14.973060 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggqrv" event={"ID":"002977b6-e9fd-4de3-8753-96a4d041c7fd","Type":"ContainerDied","Data":"2ad7de06da98f28d976dd679bc176948cb700941bd05de91282d062f6fd0786e"} Nov 24 14:21:22 crc kubenswrapper[4970]: I1124 14:21:22.037672 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggqrv" event={"ID":"002977b6-e9fd-4de3-8753-96a4d041c7fd","Type":"ContainerStarted","Data":"5cc86bf16bdc9c4bac7f7441b47fcee8eef984e66e7b6c22ebe8e62746924266"} Nov 24 14:21:23 crc kubenswrapper[4970]: I1124 14:21:23.047181 4970 generic.go:334] "Generic (PLEG): container finished" podID="002977b6-e9fd-4de3-8753-96a4d041c7fd" containerID="5cc86bf16bdc9c4bac7f7441b47fcee8eef984e66e7b6c22ebe8e62746924266" exitCode=0 Nov 24 14:21:23 crc kubenswrapper[4970]: I1124 14:21:23.047268 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggqrv" event={"ID":"002977b6-e9fd-4de3-8753-96a4d041c7fd","Type":"ContainerDied","Data":"5cc86bf16bdc9c4bac7f7441b47fcee8eef984e66e7b6c22ebe8e62746924266"} Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.267229 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.270408 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.285693 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.391565 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.391953 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4c22\" (UniqueName: \"kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.392019 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.493824 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.494143 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.494333 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4c22\" (UniqueName: \"kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.494744 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.494783 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.514991 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4c22\" (UniqueName: \"kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22\") pod \"redhat-marketplace-nj82v\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:24 crc kubenswrapper[4970]: I1124 14:21:24.592425 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:25 crc kubenswrapper[4970]: I1124 14:21:25.022047 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:25 crc kubenswrapper[4970]: W1124 14:21:25.027619 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec535443_ffb7_4bef_ae94_eb0fd0b871c3.slice/crio-8aeac3aaf202a25af6fb2b30a474fcfefb7623395fa514ab4d7ddf74d9754a2a WatchSource:0}: Error finding container 8aeac3aaf202a25af6fb2b30a474fcfefb7623395fa514ab4d7ddf74d9754a2a: Status 404 returned error can't find the container with id 8aeac3aaf202a25af6fb2b30a474fcfefb7623395fa514ab4d7ddf74d9754a2a Nov 24 14:21:25 crc kubenswrapper[4970]: I1124 14:21:25.063775 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerStarted","Data":"8aeac3aaf202a25af6fb2b30a474fcfefb7623395fa514ab4d7ddf74d9754a2a"} Nov 24 14:21:26 crc kubenswrapper[4970]: I1124 14:21:26.084370 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ggqrv" event={"ID":"002977b6-e9fd-4de3-8753-96a4d041c7fd","Type":"ContainerStarted","Data":"5661109c66334a93316d85ec4d517ec2329be5d3ea16fc52eff6038371830b87"} Nov 24 14:21:26 crc kubenswrapper[4970]: I1124 14:21:26.088043 4970 generic.go:334] "Generic (PLEG): container finished" podID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerID="c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db" exitCode=0 Nov 24 14:21:26 crc kubenswrapper[4970]: I1124 14:21:26.088097 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerDied","Data":"c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db"} Nov 24 14:21:26 crc kubenswrapper[4970]: I1124 14:21:26.120396 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ggqrv" podStartSLOduration=3.556241108 podStartE2EDuration="14.120375585s" podCreationTimestamp="2025-11-24 14:21:12 +0000 UTC" firstStartedPulling="2025-11-24 14:21:14.974914149 +0000 UTC m=+2090.262671442" lastFinishedPulling="2025-11-24 14:21:25.539048616 +0000 UTC m=+2100.826805919" observedRunningTime="2025-11-24 14:21:26.110128444 +0000 UTC m=+2101.397885767" watchObservedRunningTime="2025-11-24 14:21:26.120375585 +0000 UTC m=+2101.408132888" Nov 24 14:21:28 crc kubenswrapper[4970]: I1124 14:21:28.107676 4970 generic.go:334] "Generic (PLEG): container finished" podID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerID="9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e" exitCode=0 Nov 24 14:21:28 crc kubenswrapper[4970]: I1124 14:21:28.107781 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerDied","Data":"9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e"} Nov 24 14:21:29 crc kubenswrapper[4970]: I1124 14:21:29.117841 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerStarted","Data":"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db"} Nov 24 14:21:29 crc kubenswrapper[4970]: I1124 14:21:29.138989 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nj82v" podStartSLOduration=2.744454977 podStartE2EDuration="5.138974872s" podCreationTimestamp="2025-11-24 14:21:24 +0000 UTC" firstStartedPulling="2025-11-24 14:21:26.090154937 +0000 UTC m=+2101.377912230" lastFinishedPulling="2025-11-24 14:21:28.484674832 +0000 UTC m=+2103.772432125" observedRunningTime="2025-11-24 14:21:29.136834913 +0000 UTC m=+2104.424592216" watchObservedRunningTime="2025-11-24 14:21:29.138974872 +0000 UTC m=+2104.426732165" Nov 24 14:21:33 crc kubenswrapper[4970]: I1124 14:21:33.311433 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:33 crc kubenswrapper[4970]: I1124 14:21:33.311987 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:33 crc kubenswrapper[4970]: I1124 14:21:33.377173 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.228668 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ggqrv" Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.300414 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ggqrv"] Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.351712 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.352003 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t5s8z" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="registry-server" containerID="cri-o://f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961" gracePeriod=2 Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.592685 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.593806 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.646544 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:34 crc kubenswrapper[4970]: I1124 14:21:34.853243 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.007446 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content\") pod \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.007560 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities\") pod \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.007653 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l998g\" (UniqueName: \"kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g\") pod \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\" (UID: \"7015e74a-a3c2-4d9e-928b-ae37748bc3f8\") " Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.008086 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities" (OuterVolumeSpecName: "utilities") pod "7015e74a-a3c2-4d9e-928b-ae37748bc3f8" (UID: "7015e74a-a3c2-4d9e-928b-ae37748bc3f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.015022 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g" (OuterVolumeSpecName: "kube-api-access-l998g") pod "7015e74a-a3c2-4d9e-928b-ae37748bc3f8" (UID: "7015e74a-a3c2-4d9e-928b-ae37748bc3f8"). InnerVolumeSpecName "kube-api-access-l998g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.090590 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7015e74a-a3c2-4d9e-928b-ae37748bc3f8" (UID: "7015e74a-a3c2-4d9e-928b-ae37748bc3f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.109708 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.109747 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.109758 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l998g\" (UniqueName: \"kubernetes.io/projected/7015e74a-a3c2-4d9e-928b-ae37748bc3f8-kube-api-access-l998g\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.177030 4970 generic.go:334] "Generic (PLEG): container finished" podID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerID="f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961" exitCode=0 Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.178258 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t5s8z" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.178722 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerDied","Data":"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961"} Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.178780 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t5s8z" event={"ID":"7015e74a-a3c2-4d9e-928b-ae37748bc3f8","Type":"ContainerDied","Data":"bd10b4c54ff7d5b39db1f7df3397f7bce093948911acdd2b1bb99fdc15bf8263"} Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.178800 4970 scope.go:117] "RemoveContainer" containerID="f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.202290 4970 scope.go:117] "RemoveContainer" containerID="2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.216156 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.225840 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t5s8z"] Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.230828 4970 scope.go:117] "RemoveContainer" containerID="91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.231416 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.264904 4970 scope.go:117] "RemoveContainer" containerID="f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961" Nov 24 14:21:35 crc kubenswrapper[4970]: E1124 14:21:35.265496 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961\": container with ID starting with f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961 not found: ID does not exist" containerID="f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.265536 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961"} err="failed to get container status \"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961\": rpc error: code = NotFound desc = could not find container \"f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961\": container with ID starting with f716ef4cee1e6c2cb5c11433b5be88ac91087e52aff7c1fccc6ba67fd0dd5961 not found: ID does not exist" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.265586 4970 scope.go:117] "RemoveContainer" containerID="2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f" Nov 24 14:21:35 crc kubenswrapper[4970]: E1124 14:21:35.266135 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f\": container with ID starting with 2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f not found: ID does not exist" containerID="2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.266185 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f"} err="failed to get container status \"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f\": rpc error: code = NotFound desc = could not find container \"2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f\": container with ID starting with 2806f71972ace7999604ce2abb495a92c96306fd34759016f5347fda7f24497f not found: ID does not exist" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.266245 4970 scope.go:117] "RemoveContainer" containerID="91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4" Nov 24 14:21:35 crc kubenswrapper[4970]: E1124 14:21:35.269871 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4\": container with ID starting with 91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4 not found: ID does not exist" containerID="91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.269926 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4"} err="failed to get container status \"91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4\": rpc error: code = NotFound desc = could not find container \"91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4\": container with ID starting with 91fb7320da6f40f3dc5ca45fc69901197d2f9838c3d8dbe6c02bf0455c48d7a4 not found: ID does not exist" Nov 24 14:21:35 crc kubenswrapper[4970]: I1124 14:21:35.484853 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" path="/var/lib/kubelet/pods/7015e74a-a3c2-4d9e-928b-ae37748bc3f8/volumes" Nov 24 14:21:37 crc kubenswrapper[4970]: I1124 14:21:37.047687 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.202973 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nj82v" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="registry-server" containerID="cri-o://5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db" gracePeriod=2 Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.678179 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.778884 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4c22\" (UniqueName: \"kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22\") pod \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.779063 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content\") pod \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.779128 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities\") pod \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\" (UID: \"ec535443-ffb7-4bef-ae94-eb0fd0b871c3\") " Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.780157 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities" (OuterVolumeSpecName: "utilities") pod "ec535443-ffb7-4bef-ae94-eb0fd0b871c3" (UID: "ec535443-ffb7-4bef-ae94-eb0fd0b871c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.785492 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22" (OuterVolumeSpecName: "kube-api-access-q4c22") pod "ec535443-ffb7-4bef-ae94-eb0fd0b871c3" (UID: "ec535443-ffb7-4bef-ae94-eb0fd0b871c3"). InnerVolumeSpecName "kube-api-access-q4c22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.795609 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec535443-ffb7-4bef-ae94-eb0fd0b871c3" (UID: "ec535443-ffb7-4bef-ae94-eb0fd0b871c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.881987 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.882023 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4c22\" (UniqueName: \"kubernetes.io/projected/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-kube-api-access-q4c22\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:38 crc kubenswrapper[4970]: I1124 14:21:38.882036 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec535443-ffb7-4bef-ae94-eb0fd0b871c3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.213793 4970 generic.go:334] "Generic (PLEG): container finished" podID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerID="5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db" exitCode=0 Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.213857 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nj82v" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.213878 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerDied","Data":"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db"} Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.214911 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nj82v" event={"ID":"ec535443-ffb7-4bef-ae94-eb0fd0b871c3","Type":"ContainerDied","Data":"8aeac3aaf202a25af6fb2b30a474fcfefb7623395fa514ab4d7ddf74d9754a2a"} Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.214933 4970 scope.go:117] "RemoveContainer" containerID="5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.237063 4970 scope.go:117] "RemoveContainer" containerID="9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.249739 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.257676 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nj82v"] Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.285508 4970 scope.go:117] "RemoveContainer" containerID="c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.312393 4970 scope.go:117] "RemoveContainer" containerID="5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db" Nov 24 14:21:39 crc kubenswrapper[4970]: E1124 14:21:39.312788 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db\": container with ID starting with 5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db not found: ID does not exist" containerID="5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.312845 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db"} err="failed to get container status \"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db\": rpc error: code = NotFound desc = could not find container \"5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db\": container with ID starting with 5204a72472846fcb12f5ed8a2f8487b8a4675f261011291955ad6d93e6d9a0db not found: ID does not exist" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.312866 4970 scope.go:117] "RemoveContainer" containerID="9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e" Nov 24 14:21:39 crc kubenswrapper[4970]: E1124 14:21:39.313206 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e\": container with ID starting with 9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e not found: ID does not exist" containerID="9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.313249 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e"} err="failed to get container status \"9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e\": rpc error: code = NotFound desc = could not find container \"9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e\": container with ID starting with 9e18341081a47978f373ba09eb8a2bd3157024fa178b479954f9ec7acf7b057e not found: ID does not exist" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.313275 4970 scope.go:117] "RemoveContainer" containerID="c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db" Nov 24 14:21:39 crc kubenswrapper[4970]: E1124 14:21:39.313599 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db\": container with ID starting with c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db not found: ID does not exist" containerID="c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.313627 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db"} err="failed to get container status \"c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db\": rpc error: code = NotFound desc = could not find container \"c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db\": container with ID starting with c6a05214c90143c5949ffad63d5f484a295ed1a128e374bfae8a1b9a1f93c4db not found: ID does not exist" Nov 24 14:21:39 crc kubenswrapper[4970]: I1124 14:21:39.482379 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" path="/var/lib/kubelet/pods/ec535443-ffb7-4bef-ae94-eb0fd0b871c3/volumes" Nov 24 14:21:51 crc kubenswrapper[4970]: I1124 14:21:51.336644 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dcb0227-fbb2-492f-819b-02dd5de1a728" containerID="75668996cfbded98f2cdaee223ad3ed2359882b3043f753a489f9de4e5416b4d" exitCode=0 Nov 24 14:21:51 crc kubenswrapper[4970]: I1124 14:21:51.336694 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" event={"ID":"5dcb0227-fbb2-492f-819b-02dd5de1a728","Type":"ContainerDied","Data":"75668996cfbded98f2cdaee223ad3ed2359882b3043f753a489f9de4e5416b4d"} Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.825978 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.957763 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory\") pod \"5dcb0227-fbb2-492f-819b-02dd5de1a728\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.958866 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle\") pod \"5dcb0227-fbb2-492f-819b-02dd5de1a728\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.959381 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0\") pod \"5dcb0227-fbb2-492f-819b-02dd5de1a728\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.959520 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key\") pod \"5dcb0227-fbb2-492f-819b-02dd5de1a728\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.959622 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swvvd\" (UniqueName: \"kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd\") pod \"5dcb0227-fbb2-492f-819b-02dd5de1a728\" (UID: \"5dcb0227-fbb2-492f-819b-02dd5de1a728\") " Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.964969 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5dcb0227-fbb2-492f-819b-02dd5de1a728" (UID: "5dcb0227-fbb2-492f-819b-02dd5de1a728"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.965959 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd" (OuterVolumeSpecName: "kube-api-access-swvvd") pod "5dcb0227-fbb2-492f-819b-02dd5de1a728" (UID: "5dcb0227-fbb2-492f-819b-02dd5de1a728"). InnerVolumeSpecName "kube-api-access-swvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.997517 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory" (OuterVolumeSpecName: "inventory") pod "5dcb0227-fbb2-492f-819b-02dd5de1a728" (UID: "5dcb0227-fbb2-492f-819b-02dd5de1a728"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:21:52 crc kubenswrapper[4970]: I1124 14:21:52.999081 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5dcb0227-fbb2-492f-819b-02dd5de1a728" (UID: "5dcb0227-fbb2-492f-819b-02dd5de1a728"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.015347 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "5dcb0227-fbb2-492f-819b-02dd5de1a728" (UID: "5dcb0227-fbb2-492f-819b-02dd5de1a728"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.062697 4970 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.062738 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.062796 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swvvd\" (UniqueName: \"kubernetes.io/projected/5dcb0227-fbb2-492f-819b-02dd5de1a728-kube-api-access-swvvd\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.062819 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.062837 4970 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcb0227-fbb2-492f-819b-02dd5de1a728-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.357808 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" event={"ID":"5dcb0227-fbb2-492f-819b-02dd5de1a728","Type":"ContainerDied","Data":"1ed2a744925e7f94339f4c293345f41bc4bdc850a2519ec9239de593d85f25be"} Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.357879 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ed2a744925e7f94339f4c293345f41bc4bdc850a2519ec9239de593d85f25be" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.357989 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.463688 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z"] Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464238 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464265 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464294 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="extract-content" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464302 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="extract-content" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464315 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcb0227-fbb2-492f-819b-02dd5de1a728" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464325 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcb0227-fbb2-492f-819b-02dd5de1a728" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464336 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="extract-utilities" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464344 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="extract-utilities" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464358 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464365 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464396 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="extract-utilities" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464403 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="extract-utilities" Nov 24 14:21:53 crc kubenswrapper[4970]: E1124 14:21:53.464423 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="extract-content" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464430 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="extract-content" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464690 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec535443-ffb7-4bef-ae94-eb0fd0b871c3" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464715 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="7015e74a-a3c2-4d9e-928b-ae37748bc3f8" containerName="registry-server" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.464733 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcb0227-fbb2-492f-819b-02dd5de1a728" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.465491 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.468903 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.469360 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.469449 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.469534 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.472491 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.472785 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.473204 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.485253 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z"] Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.571553 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.571708 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.571751 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.571860 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.571966 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.572361 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpvpz\" (UniqueName: \"kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.572614 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.572667 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.572711 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.674823 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpvpz\" (UniqueName: \"kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.674949 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.674977 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675018 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675104 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675163 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675192 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675222 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.675269 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.677810 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.680925 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.681268 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.681292 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.681868 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.681925 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.682373 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.683517 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.701293 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpvpz\" (UniqueName: \"kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz\") pod \"nova-edpm-deployment-openstack-edpm-ipam-xmx7z\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:53 crc kubenswrapper[4970]: I1124 14:21:53.804745 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:21:54 crc kubenswrapper[4970]: I1124 14:21:54.327605 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z"] Nov 24 14:21:54 crc kubenswrapper[4970]: W1124 14:21:54.337011 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfcd6f60_0b2f_4066_bb72_9634b21eb168.slice/crio-8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a WatchSource:0}: Error finding container 8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a: Status 404 returned error can't find the container with id 8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a Nov 24 14:21:54 crc kubenswrapper[4970]: I1124 14:21:54.368269 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" event={"ID":"cfcd6f60-0b2f-4066-bb72-9634b21eb168","Type":"ContainerStarted","Data":"8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a"} Nov 24 14:21:55 crc kubenswrapper[4970]: I1124 14:21:55.379754 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" event={"ID":"cfcd6f60-0b2f-4066-bb72-9634b21eb168","Type":"ContainerStarted","Data":"6b50c867bfc2d4ae9790938d9fc4adcb6d8a682ce39b425928db0cdfd3eaa8b3"} Nov 24 14:21:55 crc kubenswrapper[4970]: I1124 14:21:55.399113 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" podStartSLOduration=1.955780941 podStartE2EDuration="2.399095335s" podCreationTimestamp="2025-11-24 14:21:53 +0000 UTC" firstStartedPulling="2025-11-24 14:21:54.340877061 +0000 UTC m=+2129.628634354" lastFinishedPulling="2025-11-24 14:21:54.784191455 +0000 UTC m=+2130.071948748" observedRunningTime="2025-11-24 14:21:55.396797303 +0000 UTC m=+2130.684554586" watchObservedRunningTime="2025-11-24 14:21:55.399095335 +0000 UTC m=+2130.686852628" Nov 24 14:22:11 crc kubenswrapper[4970]: I1124 14:22:11.204802 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:22:11 crc kubenswrapper[4970]: I1124 14:22:11.205445 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:22:41 crc kubenswrapper[4970]: I1124 14:22:41.204442 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:22:41 crc kubenswrapper[4970]: I1124 14:22:41.205058 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:23:11 crc kubenswrapper[4970]: I1124 14:23:11.203671 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:23:11 crc kubenswrapper[4970]: I1124 14:23:11.204135 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:23:11 crc kubenswrapper[4970]: I1124 14:23:11.204184 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:23:11 crc kubenswrapper[4970]: I1124 14:23:11.204900 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:23:11 crc kubenswrapper[4970]: I1124 14:23:11.204946 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" gracePeriod=600 Nov 24 14:23:11 crc kubenswrapper[4970]: E1124 14:23:11.325608 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:23:12 crc kubenswrapper[4970]: I1124 14:23:12.115620 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" exitCode=0 Nov 24 14:23:12 crc kubenswrapper[4970]: I1124 14:23:12.115685 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35"} Nov 24 14:23:12 crc kubenswrapper[4970]: I1124 14:23:12.115983 4970 scope.go:117] "RemoveContainer" containerID="e1830cc4b238ed4821c536ffd8727397623d1d4841e75c85e6e8bdbd1067d073" Nov 24 14:23:12 crc kubenswrapper[4970]: I1124 14:23:12.117112 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:23:12 crc kubenswrapper[4970]: E1124 14:23:12.117746 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:23:26 crc kubenswrapper[4970]: I1124 14:23:26.470037 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:23:26 crc kubenswrapper[4970]: E1124 14:23:26.470871 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:23:40 crc kubenswrapper[4970]: I1124 14:23:40.470528 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:23:40 crc kubenswrapper[4970]: E1124 14:23:40.471845 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:23:53 crc kubenswrapper[4970]: I1124 14:23:53.470810 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:23:53 crc kubenswrapper[4970]: E1124 14:23:53.471505 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:24:07 crc kubenswrapper[4970]: I1124 14:24:07.470983 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:24:07 crc kubenswrapper[4970]: E1124 14:24:07.471763 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:24:19 crc kubenswrapper[4970]: I1124 14:24:19.470415 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:24:19 crc kubenswrapper[4970]: E1124 14:24:19.471278 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:24:31 crc kubenswrapper[4970]: I1124 14:24:31.915993 4970 generic.go:334] "Generic (PLEG): container finished" podID="cfcd6f60-0b2f-4066-bb72-9634b21eb168" containerID="6b50c867bfc2d4ae9790938d9fc4adcb6d8a682ce39b425928db0cdfd3eaa8b3" exitCode=0 Nov 24 14:24:31 crc kubenswrapper[4970]: I1124 14:24:31.916088 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" event={"ID":"cfcd6f60-0b2f-4066-bb72-9634b21eb168","Type":"ContainerDied","Data":"6b50c867bfc2d4ae9790938d9fc4adcb6d8a682ce39b425928db0cdfd3eaa8b3"} Nov 24 14:24:32 crc kubenswrapper[4970]: I1124 14:24:32.471225 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:24:32 crc kubenswrapper[4970]: E1124 14:24:32.471815 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.388336 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.429973 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.430074 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.430183 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.430213 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.430240 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpvpz\" (UniqueName: \"kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.430812 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.431134 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.431201 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.431242 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle\") pod \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\" (UID: \"cfcd6f60-0b2f-4066-bb72-9634b21eb168\") " Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.442866 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz" (OuterVolumeSpecName: "kube-api-access-bpvpz") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "kube-api-access-bpvpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.444478 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.457227 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.460426 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.464141 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.465973 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory" (OuterVolumeSpecName: "inventory") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.466294 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.466373 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.475820 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "cfcd6f60-0b2f-4066-bb72-9634b21eb168" (UID: "cfcd6f60-0b2f-4066-bb72-9634b21eb168"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.532936 4970 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.532980 4970 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.532993 4970 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533005 4970 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533017 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533031 4970 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533041 4970 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533052 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cfcd6f60-0b2f-4066-bb72-9634b21eb168-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.533066 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpvpz\" (UniqueName: \"kubernetes.io/projected/cfcd6f60-0b2f-4066-bb72-9634b21eb168-kube-api-access-bpvpz\") on node \"crc\" DevicePath \"\"" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.943153 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" event={"ID":"cfcd6f60-0b2f-4066-bb72-9634b21eb168","Type":"ContainerDied","Data":"8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a"} Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.943194 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b0b3f3c4ba76cc6043a24c088ee7e20ec28c8a4898471c91bc2e95188256f2a" Nov 24 14:24:33 crc kubenswrapper[4970]: I1124 14:24:33.943253 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-xmx7z" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.049506 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn"] Nov 24 14:24:34 crc kubenswrapper[4970]: E1124 14:24:34.050106 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfcd6f60-0b2f-4066-bb72-9634b21eb168" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.050127 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfcd6f60-0b2f-4066-bb72-9634b21eb168" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.050385 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfcd6f60-0b2f-4066-bb72-9634b21eb168" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.051369 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.053404 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.053937 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.054395 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.055060 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.055876 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-sfqvr" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.067128 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn"] Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.143769 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.143871 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.143931 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.144006 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.144081 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dglgj\" (UniqueName: \"kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.144104 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.144181 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245657 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245718 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245748 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245802 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245874 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dglgj\" (UniqueName: \"kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245903 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.245928 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.250720 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.250932 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.253165 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.253599 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.254839 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.255146 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.266735 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dglgj\" (UniqueName: \"kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-whzwn\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.394624 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.930912 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn"] Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.934014 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:24:34 crc kubenswrapper[4970]: I1124 14:24:34.952216 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" event={"ID":"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2","Type":"ContainerStarted","Data":"36cf0741e007cc9c93361ea23bb2cf8a1befeea12fc1fae2544a025ba4808397"} Nov 24 14:24:35 crc kubenswrapper[4970]: I1124 14:24:35.967157 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" event={"ID":"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2","Type":"ContainerStarted","Data":"6f7065dfbbfaf46beb23a48499327c8fce3a02d2f9c1560ded7d7f14295d9635"} Nov 24 14:24:35 crc kubenswrapper[4970]: I1124 14:24:35.992002 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" podStartSLOduration=1.556021299 podStartE2EDuration="1.991981275s" podCreationTimestamp="2025-11-24 14:24:34 +0000 UTC" firstStartedPulling="2025-11-24 14:24:34.933738954 +0000 UTC m=+2290.221496257" lastFinishedPulling="2025-11-24 14:24:35.36969894 +0000 UTC m=+2290.657456233" observedRunningTime="2025-11-24 14:24:35.990718716 +0000 UTC m=+2291.278476009" watchObservedRunningTime="2025-11-24 14:24:35.991981275 +0000 UTC m=+2291.279738578" Nov 24 14:24:46 crc kubenswrapper[4970]: I1124 14:24:46.470946 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:24:46 crc kubenswrapper[4970]: E1124 14:24:46.471729 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:24:58 crc kubenswrapper[4970]: I1124 14:24:58.470921 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:24:58 crc kubenswrapper[4970]: E1124 14:24:58.471853 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:25:10 crc kubenswrapper[4970]: I1124 14:25:10.470603 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:25:10 crc kubenswrapper[4970]: E1124 14:25:10.471383 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:25:23 crc kubenswrapper[4970]: I1124 14:25:23.473482 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:25:23 crc kubenswrapper[4970]: E1124 14:25:23.474930 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:25:34 crc kubenswrapper[4970]: I1124 14:25:34.470709 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:25:34 crc kubenswrapper[4970]: E1124 14:25:34.471484 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.205696 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.210416 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.214422 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.229971 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.230073 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.230161 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8wmn\" (UniqueName: \"kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.331388 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.331471 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.331521 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8wmn\" (UniqueName: \"kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.332049 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.332153 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.364474 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8wmn\" (UniqueName: \"kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn\") pod \"community-operators-nrx68\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.538044 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:38 crc kubenswrapper[4970]: I1124 14:25:38.992651 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:39 crc kubenswrapper[4970]: I1124 14:25:39.682893 4970 generic.go:334] "Generic (PLEG): container finished" podID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerID="6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344" exitCode=0 Nov 24 14:25:39 crc kubenswrapper[4970]: I1124 14:25:39.682976 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerDied","Data":"6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344"} Nov 24 14:25:39 crc kubenswrapper[4970]: I1124 14:25:39.683183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerStarted","Data":"fd71b7ecccfc29003077abd9cfcf90c817d62f58ead2b7d50bb5ecbecceb1da9"} Nov 24 14:25:41 crc kubenswrapper[4970]: I1124 14:25:41.706427 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerStarted","Data":"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59"} Nov 24 14:25:42 crc kubenswrapper[4970]: I1124 14:25:42.716121 4970 generic.go:334] "Generic (PLEG): container finished" podID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerID="005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59" exitCode=0 Nov 24 14:25:42 crc kubenswrapper[4970]: I1124 14:25:42.716185 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerDied","Data":"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59"} Nov 24 14:25:43 crc kubenswrapper[4970]: I1124 14:25:43.727131 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerStarted","Data":"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057"} Nov 24 14:25:43 crc kubenswrapper[4970]: I1124 14:25:43.762257 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nrx68" podStartSLOduration=2.326596651 podStartE2EDuration="5.762236775s" podCreationTimestamp="2025-11-24 14:25:38 +0000 UTC" firstStartedPulling="2025-11-24 14:25:39.685897214 +0000 UTC m=+2354.973654507" lastFinishedPulling="2025-11-24 14:25:43.121537328 +0000 UTC m=+2358.409294631" observedRunningTime="2025-11-24 14:25:43.758479246 +0000 UTC m=+2359.046236549" watchObservedRunningTime="2025-11-24 14:25:43.762236775 +0000 UTC m=+2359.049994068" Nov 24 14:25:46 crc kubenswrapper[4970]: I1124 14:25:46.471114 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:25:46 crc kubenswrapper[4970]: E1124 14:25:46.471531 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:25:48 crc kubenswrapper[4970]: I1124 14:25:48.539006 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:48 crc kubenswrapper[4970]: I1124 14:25:48.539529 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:48 crc kubenswrapper[4970]: I1124 14:25:48.598993 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:48 crc kubenswrapper[4970]: I1124 14:25:48.862226 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:48 crc kubenswrapper[4970]: I1124 14:25:48.936188 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:50 crc kubenswrapper[4970]: I1124 14:25:50.808040 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nrx68" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="registry-server" containerID="cri-o://27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057" gracePeriod=2 Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.264441 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.393829 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8wmn\" (UniqueName: \"kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn\") pod \"78aec434-1c84-40e1-ab30-c9df6bd1155b\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.394222 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content\") pod \"78aec434-1c84-40e1-ab30-c9df6bd1155b\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.394331 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities\") pod \"78aec434-1c84-40e1-ab30-c9df6bd1155b\" (UID: \"78aec434-1c84-40e1-ab30-c9df6bd1155b\") " Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.396189 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities" (OuterVolumeSpecName: "utilities") pod "78aec434-1c84-40e1-ab30-c9df6bd1155b" (UID: "78aec434-1c84-40e1-ab30-c9df6bd1155b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.400744 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn" (OuterVolumeSpecName: "kube-api-access-w8wmn") pod "78aec434-1c84-40e1-ab30-c9df6bd1155b" (UID: "78aec434-1c84-40e1-ab30-c9df6bd1155b"). InnerVolumeSpecName "kube-api-access-w8wmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.496656 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.496876 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8wmn\" (UniqueName: \"kubernetes.io/projected/78aec434-1c84-40e1-ab30-c9df6bd1155b-kube-api-access-w8wmn\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.820219 4970 generic.go:334] "Generic (PLEG): container finished" podID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerID="27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057" exitCode=0 Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.820268 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerDied","Data":"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057"} Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.820301 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrx68" event={"ID":"78aec434-1c84-40e1-ab30-c9df6bd1155b","Type":"ContainerDied","Data":"fd71b7ecccfc29003077abd9cfcf90c817d62f58ead2b7d50bb5ecbecceb1da9"} Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.820321 4970 scope.go:117] "RemoveContainer" containerID="27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.820343 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrx68" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.853910 4970 scope.go:117] "RemoveContainer" containerID="005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.909272 4970 scope.go:117] "RemoveContainer" containerID="6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.961665 4970 scope.go:117] "RemoveContainer" containerID="27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057" Nov 24 14:25:51 crc kubenswrapper[4970]: E1124 14:25:51.962880 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057\": container with ID starting with 27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057 not found: ID does not exist" containerID="27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.962933 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057"} err="failed to get container status \"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057\": rpc error: code = NotFound desc = could not find container \"27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057\": container with ID starting with 27c05eb97a1b3077039e8f1939f5fa7f9f972dc4818bfbbb8bffdb811b179057 not found: ID does not exist" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.962969 4970 scope.go:117] "RemoveContainer" containerID="005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59" Nov 24 14:25:51 crc kubenswrapper[4970]: E1124 14:25:51.963512 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59\": container with ID starting with 005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59 not found: ID does not exist" containerID="005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.963690 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59"} err="failed to get container status \"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59\": rpc error: code = NotFound desc = could not find container \"005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59\": container with ID starting with 005fbf7e1b4d0a4a7bcf0df556cb782af61e6caa6322f800c4ba6c9753a8aa59 not found: ID does not exist" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.963797 4970 scope.go:117] "RemoveContainer" containerID="6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344" Nov 24 14:25:51 crc kubenswrapper[4970]: E1124 14:25:51.964182 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344\": container with ID starting with 6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344 not found: ID does not exist" containerID="6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344" Nov 24 14:25:51 crc kubenswrapper[4970]: I1124 14:25:51.964218 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344"} err="failed to get container status \"6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344\": rpc error: code = NotFound desc = could not find container \"6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344\": container with ID starting with 6b91a34a1501d4c725b9252990bc655ef003fa12460ef0f5349cb07ce3b89344 not found: ID does not exist" Nov 24 14:25:52 crc kubenswrapper[4970]: I1124 14:25:52.128161 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78aec434-1c84-40e1-ab30-c9df6bd1155b" (UID: "78aec434-1c84-40e1-ab30-c9df6bd1155b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:25:52 crc kubenswrapper[4970]: I1124 14:25:52.213613 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78aec434-1c84-40e1-ab30-c9df6bd1155b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:25:52 crc kubenswrapper[4970]: I1124 14:25:52.479362 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:52 crc kubenswrapper[4970]: I1124 14:25:52.492311 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nrx68"] Nov 24 14:25:53 crc kubenswrapper[4970]: I1124 14:25:53.484161 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" path="/var/lib/kubelet/pods/78aec434-1c84-40e1-ab30-c9df6bd1155b/volumes" Nov 24 14:26:00 crc kubenswrapper[4970]: I1124 14:26:00.472383 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:26:00 crc kubenswrapper[4970]: E1124 14:26:00.473327 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:26:13 crc kubenswrapper[4970]: I1124 14:26:13.470974 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:26:13 crc kubenswrapper[4970]: E1124 14:26:13.471686 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:26:25 crc kubenswrapper[4970]: I1124 14:26:25.489347 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:26:25 crc kubenswrapper[4970]: E1124 14:26:25.490606 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:26:40 crc kubenswrapper[4970]: I1124 14:26:40.470819 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:26:40 crc kubenswrapper[4970]: E1124 14:26:40.471539 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:26:53 crc kubenswrapper[4970]: I1124 14:26:53.455825 4970 generic.go:334] "Generic (PLEG): container finished" podID="196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" containerID="6f7065dfbbfaf46beb23a48499327c8fce3a02d2f9c1560ded7d7f14295d9635" exitCode=0 Nov 24 14:26:53 crc kubenswrapper[4970]: I1124 14:26:53.455965 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" event={"ID":"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2","Type":"ContainerDied","Data":"6f7065dfbbfaf46beb23a48499327c8fce3a02d2f9c1560ded7d7f14295d9635"} Nov 24 14:26:54 crc kubenswrapper[4970]: I1124 14:26:54.471068 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:26:54 crc kubenswrapper[4970]: E1124 14:26:54.471743 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:26:54 crc kubenswrapper[4970]: I1124 14:26:54.897848 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.027925 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dglgj\" (UniqueName: \"kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028050 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028080 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028181 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028294 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028341 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.028378 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle\") pod \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\" (UID: \"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2\") " Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.033882 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj" (OuterVolumeSpecName: "kube-api-access-dglgj") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "kube-api-access-dglgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.036192 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.054624 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.058413 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.064126 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory" (OuterVolumeSpecName: "inventory") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.077558 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.085156 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" (UID: "196cf6ec-261f-4c0e-87f4-30c0e60e7bd2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130454 4970 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130520 4970 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130533 4970 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130544 4970 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130632 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dglgj\" (UniqueName: \"kubernetes.io/projected/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-kube-api-access-dglgj\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130649 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.130685 4970 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/196cf6ec-261f-4c0e-87f4-30c0e60e7bd2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.486365 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.494732 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-whzwn" event={"ID":"196cf6ec-261f-4c0e-87f4-30c0e60e7bd2","Type":"ContainerDied","Data":"36cf0741e007cc9c93361ea23bb2cf8a1befeea12fc1fae2544a025ba4808397"} Nov 24 14:26:55 crc kubenswrapper[4970]: I1124 14:26:55.494792 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36cf0741e007cc9c93361ea23bb2cf8a1befeea12fc1fae2544a025ba4808397" Nov 24 14:27:05 crc kubenswrapper[4970]: I1124 14:27:05.484072 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:27:05 crc kubenswrapper[4970]: E1124 14:27:05.484951 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:27:19 crc kubenswrapper[4970]: I1124 14:27:19.470851 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:27:19 crc kubenswrapper[4970]: E1124 14:27:19.471726 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:27:34 crc kubenswrapper[4970]: I1124 14:27:34.470972 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:27:34 crc kubenswrapper[4970]: E1124 14:27:34.472015 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:27:45 crc kubenswrapper[4970]: I1124 14:27:45.484303 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:27:45 crc kubenswrapper[4970]: E1124 14:27:45.485167 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:27:46 crc kubenswrapper[4970]: E1124 14:27:46.284891 4970 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.12:36058->38.102.83.12:33579: write tcp 38.102.83.12:36058->38.102.83.12:33579: write: broken pipe Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.782456 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:27:54 crc kubenswrapper[4970]: E1124 14:27:54.783460 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="registry-server" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783478 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="registry-server" Nov 24 14:27:54 crc kubenswrapper[4970]: E1124 14:27:54.783495 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783504 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:27:54 crc kubenswrapper[4970]: E1124 14:27:54.783518 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="extract-utilities" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783527 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="extract-utilities" Nov 24 14:27:54 crc kubenswrapper[4970]: E1124 14:27:54.783549 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="extract-content" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783556 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="extract-content" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783819 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="78aec434-1c84-40e1-ab30-c9df6bd1155b" containerName="registry-server" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.783835 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="196cf6ec-261f-4c0e-87f4-30c0e60e7bd2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.784721 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.786997 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.787332 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2qq2r" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.787357 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.787527 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.801604 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936185 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936233 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936295 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936342 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936372 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936399 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936428 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmlwl\" (UniqueName: \"kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936520 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:54 crc kubenswrapper[4970]: I1124 14:27:54.936560 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038077 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038169 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038206 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038287 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038312 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038356 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038402 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038461 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmlwl\" (UniqueName: \"kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038513 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038823 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.038973 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.039219 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.039408 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.040036 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.046715 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.047329 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.048084 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.074564 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmlwl\" (UniqueName: \"kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.092103 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.105818 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:27:55 crc kubenswrapper[4970]: I1124 14:27:55.630959 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 24 14:27:56 crc kubenswrapper[4970]: I1124 14:27:56.337001 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1ffd9b46-906d-47a2-870e-6ad3d5381c2d","Type":"ContainerStarted","Data":"ffd8964ff7f85fa42f4ce7c1fa4e4456006fb7dc62f8c3d071ecf120f8550030"} Nov 24 14:27:58 crc kubenswrapper[4970]: I1124 14:27:58.470867 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:27:58 crc kubenswrapper[4970]: E1124 14:27:58.471514 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:28:12 crc kubenswrapper[4970]: I1124 14:28:12.470977 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:28:24 crc kubenswrapper[4970]: E1124 14:28:24.683677 4970 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 24 14:28:24 crc kubenswrapper[4970]: E1124 14:28:24.684410 4970 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmlwl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(1ffd9b46-906d-47a2-870e-6ad3d5381c2d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 14:28:24 crc kubenswrapper[4970]: E1124 14:28:24.685841 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" Nov 24 14:28:25 crc kubenswrapper[4970]: I1124 14:28:25.693913 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c"} Nov 24 14:28:25 crc kubenswrapper[4970]: E1124 14:28:25.695806 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" Nov 24 14:28:40 crc kubenswrapper[4970]: I1124 14:28:40.942788 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 24 14:28:41 crc kubenswrapper[4970]: I1124 14:28:41.973474 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1ffd9b46-906d-47a2-870e-6ad3d5381c2d","Type":"ContainerStarted","Data":"9aa63fc36451f266eb948fc187fa0d420c264fc9568c5b18a00ba6765605617a"} Nov 24 14:28:41 crc kubenswrapper[4970]: I1124 14:28:41.992129 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.693562984 podStartE2EDuration="48.992103174s" podCreationTimestamp="2025-11-24 14:27:53 +0000 UTC" firstStartedPulling="2025-11-24 14:27:55.640889841 +0000 UTC m=+2490.928647174" lastFinishedPulling="2025-11-24 14:28:40.939430071 +0000 UTC m=+2536.227187364" observedRunningTime="2025-11-24 14:28:41.986721179 +0000 UTC m=+2537.274478482" watchObservedRunningTime="2025-11-24 14:28:41.992103174 +0000 UTC m=+2537.279860467" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.168932 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n"] Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.181336 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.184228 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.185901 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.191880 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n"] Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.313243 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jz4\" (UniqueName: \"kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.313628 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.313702 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.414755 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jz4\" (UniqueName: \"kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.414796 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.414846 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.415719 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.421749 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.435073 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jz4\" (UniqueName: \"kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4\") pod \"collect-profiles-29399910-22t7n\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.513101 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:00 crc kubenswrapper[4970]: I1124 14:30:00.979263 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n"] Nov 24 14:30:01 crc kubenswrapper[4970]: I1124 14:30:01.787698 4970 generic.go:334] "Generic (PLEG): container finished" podID="52d239e5-8e43-4745-bb85-872560a2c15e" containerID="a0fff36b6e4eb611e5028a0d946f2508e901c9b0e7bdf43ef8bdbef3bfbbbfc0" exitCode=0 Nov 24 14:30:01 crc kubenswrapper[4970]: I1124 14:30:01.787770 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" event={"ID":"52d239e5-8e43-4745-bb85-872560a2c15e","Type":"ContainerDied","Data":"a0fff36b6e4eb611e5028a0d946f2508e901c9b0e7bdf43ef8bdbef3bfbbbfc0"} Nov 24 14:30:01 crc kubenswrapper[4970]: I1124 14:30:01.787959 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" event={"ID":"52d239e5-8e43-4745-bb85-872560a2c15e","Type":"ContainerStarted","Data":"f4728778afdb7e1d4cd3c8cc5c218152e23a4a21dd7fbaa7be6a203ae1266429"} Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.253844 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.370126 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume\") pod \"52d239e5-8e43-4745-bb85-872560a2c15e\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.370210 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8jz4\" (UniqueName: \"kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4\") pod \"52d239e5-8e43-4745-bb85-872560a2c15e\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.370369 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume\") pod \"52d239e5-8e43-4745-bb85-872560a2c15e\" (UID: \"52d239e5-8e43-4745-bb85-872560a2c15e\") " Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.370896 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume" (OuterVolumeSpecName: "config-volume") pod "52d239e5-8e43-4745-bb85-872560a2c15e" (UID: "52d239e5-8e43-4745-bb85-872560a2c15e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.375767 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52d239e5-8e43-4745-bb85-872560a2c15e" (UID: "52d239e5-8e43-4745-bb85-872560a2c15e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.383727 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4" (OuterVolumeSpecName: "kube-api-access-g8jz4") pod "52d239e5-8e43-4745-bb85-872560a2c15e" (UID: "52d239e5-8e43-4745-bb85-872560a2c15e"). InnerVolumeSpecName "kube-api-access-g8jz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.472548 4970 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52d239e5-8e43-4745-bb85-872560a2c15e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.472679 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52d239e5-8e43-4745-bb85-872560a2c15e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.472704 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8jz4\" (UniqueName: \"kubernetes.io/projected/52d239e5-8e43-4745-bb85-872560a2c15e-kube-api-access-g8jz4\") on node \"crc\" DevicePath \"\"" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.825767 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" event={"ID":"52d239e5-8e43-4745-bb85-872560a2c15e","Type":"ContainerDied","Data":"f4728778afdb7e1d4cd3c8cc5c218152e23a4a21dd7fbaa7be6a203ae1266429"} Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.825812 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4728778afdb7e1d4cd3c8cc5c218152e23a4a21dd7fbaa7be6a203ae1266429" Nov 24 14:30:03 crc kubenswrapper[4970]: I1124 14:30:03.825899 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399910-22t7n" Nov 24 14:30:04 crc kubenswrapper[4970]: I1124 14:30:04.335004 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4"] Nov 24 14:30:04 crc kubenswrapper[4970]: I1124 14:30:04.342166 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399865-tr7d4"] Nov 24 14:30:05 crc kubenswrapper[4970]: I1124 14:30:05.486077 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b7c8f1-519f-4063-84e6-4e0b85fe57f9" path="/var/lib/kubelet/pods/53b7c8f1-519f-4063-84e6-4e0b85fe57f9/volumes" Nov 24 14:30:41 crc kubenswrapper[4970]: I1124 14:30:41.204614 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:30:41 crc kubenswrapper[4970]: I1124 14:30:41.205234 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:30:47 crc kubenswrapper[4970]: I1124 14:30:47.260318 4970 scope.go:117] "RemoveContainer" containerID="6fce3202ec2e0caf65c347c2f8362a4b6ef43ddb43dc94cbefa1117c443b4b12" Nov 24 14:31:11 crc kubenswrapper[4970]: I1124 14:31:11.204240 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:31:11 crc kubenswrapper[4970]: I1124 14:31:11.204934 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.029547 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:13 crc kubenswrapper[4970]: E1124 14:31:13.030256 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52d239e5-8e43-4745-bb85-872560a2c15e" containerName="collect-profiles" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.030270 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="52d239e5-8e43-4745-bb85-872560a2c15e" containerName="collect-profiles" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.030538 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="52d239e5-8e43-4745-bb85-872560a2c15e" containerName="collect-profiles" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.032317 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.039542 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.111736 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9s9t\" (UniqueName: \"kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.112041 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.112070 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.213949 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9s9t\" (UniqueName: \"kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.214039 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.214078 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.214632 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.214682 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.234121 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9s9t\" (UniqueName: \"kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t\") pod \"certified-operators-d4fmm\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.386995 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:13 crc kubenswrapper[4970]: I1124 14:31:13.860871 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:14 crc kubenswrapper[4970]: I1124 14:31:14.605398 4970 generic.go:334] "Generic (PLEG): container finished" podID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerID="487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458" exitCode=0 Nov 24 14:31:14 crc kubenswrapper[4970]: I1124 14:31:14.605519 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerDied","Data":"487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458"} Nov 24 14:31:14 crc kubenswrapper[4970]: I1124 14:31:14.606068 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerStarted","Data":"f971994bd7dce8febaea8106a89cc199dc75fc09a7ad7aeab3ebb8b685237338"} Nov 24 14:31:14 crc kubenswrapper[4970]: I1124 14:31:14.607774 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:31:16 crc kubenswrapper[4970]: I1124 14:31:16.631945 4970 generic.go:334] "Generic (PLEG): container finished" podID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerID="7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4" exitCode=0 Nov 24 14:31:16 crc kubenswrapper[4970]: I1124 14:31:16.631998 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerDied","Data":"7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4"} Nov 24 14:31:17 crc kubenswrapper[4970]: I1124 14:31:17.641405 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerStarted","Data":"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf"} Nov 24 14:31:17 crc kubenswrapper[4970]: I1124 14:31:17.661100 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d4fmm" podStartSLOduration=2.035138658 podStartE2EDuration="4.661083791s" podCreationTimestamp="2025-11-24 14:31:13 +0000 UTC" firstStartedPulling="2025-11-24 14:31:14.607363102 +0000 UTC m=+2689.895120415" lastFinishedPulling="2025-11-24 14:31:17.233308255 +0000 UTC m=+2692.521065548" observedRunningTime="2025-11-24 14:31:17.658743377 +0000 UTC m=+2692.946500700" watchObservedRunningTime="2025-11-24 14:31:17.661083791 +0000 UTC m=+2692.948841074" Nov 24 14:31:23 crc kubenswrapper[4970]: I1124 14:31:23.387446 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:23 crc kubenswrapper[4970]: I1124 14:31:23.388275 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:23 crc kubenswrapper[4970]: I1124 14:31:23.444097 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:23 crc kubenswrapper[4970]: I1124 14:31:23.766789 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:23 crc kubenswrapper[4970]: I1124 14:31:23.818275 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:25 crc kubenswrapper[4970]: I1124 14:31:25.731609 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d4fmm" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="registry-server" containerID="cri-o://e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf" gracePeriod=2 Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.278673 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.377650 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities\") pod \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.377838 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content\") pod \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.377939 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9s9t\" (UniqueName: \"kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t\") pod \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\" (UID: \"b00bb841-e106-4e7f-b6b8-c0673eb1214c\") " Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.379106 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities" (OuterVolumeSpecName: "utilities") pod "b00bb841-e106-4e7f-b6b8-c0673eb1214c" (UID: "b00bb841-e106-4e7f-b6b8-c0673eb1214c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.387503 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t" (OuterVolumeSpecName: "kube-api-access-k9s9t") pod "b00bb841-e106-4e7f-b6b8-c0673eb1214c" (UID: "b00bb841-e106-4e7f-b6b8-c0673eb1214c"). InnerVolumeSpecName "kube-api-access-k9s9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.481144 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.481226 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9s9t\" (UniqueName: \"kubernetes.io/projected/b00bb841-e106-4e7f-b6b8-c0673eb1214c-kube-api-access-k9s9t\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.514039 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b00bb841-e106-4e7f-b6b8-c0673eb1214c" (UID: "b00bb841-e106-4e7f-b6b8-c0673eb1214c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.583699 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b00bb841-e106-4e7f-b6b8-c0673eb1214c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.742822 4970 generic.go:334] "Generic (PLEG): container finished" podID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerID="e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf" exitCode=0 Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.742859 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerDied","Data":"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf"} Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.742884 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4fmm" event={"ID":"b00bb841-e106-4e7f-b6b8-c0673eb1214c","Type":"ContainerDied","Data":"f971994bd7dce8febaea8106a89cc199dc75fc09a7ad7aeab3ebb8b685237338"} Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.742901 4970 scope.go:117] "RemoveContainer" containerID="e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.743004 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4fmm" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.781645 4970 scope.go:117] "RemoveContainer" containerID="7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.792242 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.800993 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d4fmm"] Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.812706 4970 scope.go:117] "RemoveContainer" containerID="487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.865928 4970 scope.go:117] "RemoveContainer" containerID="e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf" Nov 24 14:31:26 crc kubenswrapper[4970]: E1124 14:31:26.866445 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf\": container with ID starting with e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf not found: ID does not exist" containerID="e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.866506 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf"} err="failed to get container status \"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf\": rpc error: code = NotFound desc = could not find container \"e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf\": container with ID starting with e1cd19a0a0cb44b97855cf66f05886061f6687c5f74bccd985bd0b1f19447abf not found: ID does not exist" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.866541 4970 scope.go:117] "RemoveContainer" containerID="7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4" Nov 24 14:31:26 crc kubenswrapper[4970]: E1124 14:31:26.867569 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4\": container with ID starting with 7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4 not found: ID does not exist" containerID="7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.867646 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4"} err="failed to get container status \"7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4\": rpc error: code = NotFound desc = could not find container \"7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4\": container with ID starting with 7a3fb9f238adea3e886039b5d819385af0951e749e4cbe2c1dc0500c82d443a4 not found: ID does not exist" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.867680 4970 scope.go:117] "RemoveContainer" containerID="487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458" Nov 24 14:31:26 crc kubenswrapper[4970]: E1124 14:31:26.868074 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458\": container with ID starting with 487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458 not found: ID does not exist" containerID="487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458" Nov 24 14:31:26 crc kubenswrapper[4970]: I1124 14:31:26.868122 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458"} err="failed to get container status \"487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458\": rpc error: code = NotFound desc = could not find container \"487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458\": container with ID starting with 487e91269399dc996fd44dedd0282216e77ec7ba6bbf6bc27a12db50836d5458 not found: ID does not exist" Nov 24 14:31:27 crc kubenswrapper[4970]: I1124 14:31:27.504122 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" path="/var/lib/kubelet/pods/b00bb841-e106-4e7f-b6b8-c0673eb1214c/volumes" Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.204069 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.204647 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.204722 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.205689 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.205758 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c" gracePeriod=600 Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.908612 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c" exitCode=0 Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.909312 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c"} Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.909364 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709"} Nov 24 14:31:41 crc kubenswrapper[4970]: I1124 14:31:41.909386 4970 scope.go:117] "RemoveContainer" containerID="5ce11ae90104688ca9f6106e6be1e5471aeb7e5be081eded5cc26e223267de35" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.535463 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:31:55 crc kubenswrapper[4970]: E1124 14:31:55.536366 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="registry-server" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.536382 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="registry-server" Nov 24 14:31:55 crc kubenswrapper[4970]: E1124 14:31:55.536409 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="extract-content" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.536418 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="extract-content" Nov 24 14:31:55 crc kubenswrapper[4970]: E1124 14:31:55.536435 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="extract-utilities" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.536445 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="extract-utilities" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.536711 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b00bb841-e106-4e7f-b6b8-c0673eb1214c" containerName="registry-server" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.538247 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.553515 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.719304 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.719463 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6ngx\" (UniqueName: \"kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.719631 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.821408 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6ngx\" (UniqueName: \"kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.821652 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.821684 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.822280 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.822967 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.854415 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6ngx\" (UniqueName: \"kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx\") pod \"redhat-marketplace-9xqjr\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:55 crc kubenswrapper[4970]: I1124 14:31:55.863928 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:31:56 crc kubenswrapper[4970]: I1124 14:31:56.463569 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:31:57 crc kubenswrapper[4970]: I1124 14:31:57.116488 4970 generic.go:334] "Generic (PLEG): container finished" podID="707341a3-d708-401e-8527-26947cbf3d30" containerID="d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c" exitCode=0 Nov 24 14:31:57 crc kubenswrapper[4970]: I1124 14:31:57.116536 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerDied","Data":"d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c"} Nov 24 14:31:57 crc kubenswrapper[4970]: I1124 14:31:57.116566 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerStarted","Data":"e9954a4303bc3adb9b0c13a852c2a5359b83dfe18c37bcd2746245ee1a98a218"} Nov 24 14:31:58 crc kubenswrapper[4970]: I1124 14:31:58.126134 4970 generic.go:334] "Generic (PLEG): container finished" podID="707341a3-d708-401e-8527-26947cbf3d30" containerID="9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde" exitCode=0 Nov 24 14:31:58 crc kubenswrapper[4970]: I1124 14:31:58.126218 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerDied","Data":"9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde"} Nov 24 14:32:00 crc kubenswrapper[4970]: I1124 14:32:00.152123 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerStarted","Data":"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc"} Nov 24 14:32:00 crc kubenswrapper[4970]: I1124 14:32:00.174202 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9xqjr" podStartSLOduration=3.202657883 podStartE2EDuration="5.174183726s" podCreationTimestamp="2025-11-24 14:31:55 +0000 UTC" firstStartedPulling="2025-11-24 14:31:57.1190731 +0000 UTC m=+2732.406830393" lastFinishedPulling="2025-11-24 14:31:59.090598943 +0000 UTC m=+2734.378356236" observedRunningTime="2025-11-24 14:32:00.171936157 +0000 UTC m=+2735.459693450" watchObservedRunningTime="2025-11-24 14:32:00.174183726 +0000 UTC m=+2735.461941019" Nov 24 14:32:05 crc kubenswrapper[4970]: I1124 14:32:05.864700 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:05 crc kubenswrapper[4970]: I1124 14:32:05.865488 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:05 crc kubenswrapper[4970]: I1124 14:32:05.936138 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:06 crc kubenswrapper[4970]: I1124 14:32:06.313188 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:06 crc kubenswrapper[4970]: I1124 14:32:06.378217 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.239176 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9xqjr" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="registry-server" containerID="cri-o://16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc" gracePeriod=2 Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.696409 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.864950 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6ngx\" (UniqueName: \"kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx\") pod \"707341a3-d708-401e-8527-26947cbf3d30\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.865338 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities\") pod \"707341a3-d708-401e-8527-26947cbf3d30\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.865360 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content\") pod \"707341a3-d708-401e-8527-26947cbf3d30\" (UID: \"707341a3-d708-401e-8527-26947cbf3d30\") " Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.866164 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities" (OuterVolumeSpecName: "utilities") pod "707341a3-d708-401e-8527-26947cbf3d30" (UID: "707341a3-d708-401e-8527-26947cbf3d30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.870830 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx" (OuterVolumeSpecName: "kube-api-access-z6ngx") pod "707341a3-d708-401e-8527-26947cbf3d30" (UID: "707341a3-d708-401e-8527-26947cbf3d30"). InnerVolumeSpecName "kube-api-access-z6ngx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.883457 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "707341a3-d708-401e-8527-26947cbf3d30" (UID: "707341a3-d708-401e-8527-26947cbf3d30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.967482 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6ngx\" (UniqueName: \"kubernetes.io/projected/707341a3-d708-401e-8527-26947cbf3d30-kube-api-access-z6ngx\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.967524 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:08 crc kubenswrapper[4970]: I1124 14:32:08.967539 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/707341a3-d708-401e-8527-26947cbf3d30-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.253820 4970 generic.go:334] "Generic (PLEG): container finished" podID="707341a3-d708-401e-8527-26947cbf3d30" containerID="16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc" exitCode=0 Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.253889 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerDied","Data":"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc"} Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.253929 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9xqjr" event={"ID":"707341a3-d708-401e-8527-26947cbf3d30","Type":"ContainerDied","Data":"e9954a4303bc3adb9b0c13a852c2a5359b83dfe18c37bcd2746245ee1a98a218"} Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.253957 4970 scope.go:117] "RemoveContainer" containerID="16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.254169 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9xqjr" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.289037 4970 scope.go:117] "RemoveContainer" containerID="9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.332229 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.339743 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9xqjr"] Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.360363 4970 scope.go:117] "RemoveContainer" containerID="d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.395354 4970 scope.go:117] "RemoveContainer" containerID="16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc" Nov 24 14:32:09 crc kubenswrapper[4970]: E1124 14:32:09.400119 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc\": container with ID starting with 16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc not found: ID does not exist" containerID="16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.400172 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc"} err="failed to get container status \"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc\": rpc error: code = NotFound desc = could not find container \"16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc\": container with ID starting with 16aae6b2d5e62375550c2b1aa292db93652dfc5f7347b8ea98ab6cb99644eecc not found: ID does not exist" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.400197 4970 scope.go:117] "RemoveContainer" containerID="9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde" Nov 24 14:32:09 crc kubenswrapper[4970]: E1124 14:32:09.400537 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde\": container with ID starting with 9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde not found: ID does not exist" containerID="9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.400558 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde"} err="failed to get container status \"9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde\": rpc error: code = NotFound desc = could not find container \"9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde\": container with ID starting with 9fdbbf53be3e7ab0ac487ed94ab0d87458be08d11b9d6f0e88b5361d4b1f5fde not found: ID does not exist" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.400571 4970 scope.go:117] "RemoveContainer" containerID="d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c" Nov 24 14:32:09 crc kubenswrapper[4970]: E1124 14:32:09.400990 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c\": container with ID starting with d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c not found: ID does not exist" containerID="d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.401018 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c"} err="failed to get container status \"d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c\": rpc error: code = NotFound desc = could not find container \"d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c\": container with ID starting with d9d5b515248b3e5c862103807e5aa9b1ed2d12da4be8d17e738e671de71ecb7c not found: ID does not exist" Nov 24 14:32:09 crc kubenswrapper[4970]: I1124 14:32:09.480820 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="707341a3-d708-401e-8527-26947cbf3d30" path="/var/lib/kubelet/pods/707341a3-d708-401e-8527-26947cbf3d30/volumes" Nov 24 14:33:41 crc kubenswrapper[4970]: I1124 14:33:41.204924 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:33:41 crc kubenswrapper[4970]: I1124 14:33:41.205849 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:34:11 crc kubenswrapper[4970]: I1124 14:34:11.204512 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:34:11 crc kubenswrapper[4970]: I1124 14:34:11.205268 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.203962 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.204504 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.204553 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.205249 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.205302 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" gracePeriod=600 Nov 24 14:34:41 crc kubenswrapper[4970]: E1124 14:34:41.347800 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.844331 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" exitCode=0 Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.844458 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709"} Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.844696 4970 scope.go:117] "RemoveContainer" containerID="8183bbbac50ade7c3ca00d8aae2518e36dee27865cc01c67b091ed3da3387d0c" Nov 24 14:34:41 crc kubenswrapper[4970]: I1124 14:34:41.845642 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:34:41 crc kubenswrapper[4970]: E1124 14:34:41.846195 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:34:56 crc kubenswrapper[4970]: I1124 14:34:56.470838 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:34:56 crc kubenswrapper[4970]: E1124 14:34:56.471849 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:35:09 crc kubenswrapper[4970]: I1124 14:35:09.471372 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:35:09 crc kubenswrapper[4970]: E1124 14:35:09.474152 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:35:22 crc kubenswrapper[4970]: I1124 14:35:22.471459 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:35:22 crc kubenswrapper[4970]: E1124 14:35:22.472539 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:35:34 crc kubenswrapper[4970]: I1124 14:35:34.470736 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:35:34 crc kubenswrapper[4970]: E1124 14:35:34.471685 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:35:48 crc kubenswrapper[4970]: I1124 14:35:48.470198 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:35:48 crc kubenswrapper[4970]: E1124 14:35:48.470976 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:36:02 crc kubenswrapper[4970]: I1124 14:36:02.471445 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:36:02 crc kubenswrapper[4970]: E1124 14:36:02.472872 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:36:17 crc kubenswrapper[4970]: I1124 14:36:17.476953 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:36:17 crc kubenswrapper[4970]: E1124 14:36:17.477853 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:36:28 crc kubenswrapper[4970]: I1124 14:36:28.470086 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:36:28 crc kubenswrapper[4970]: E1124 14:36:28.470885 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:36:43 crc kubenswrapper[4970]: I1124 14:36:43.471507 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:36:43 crc kubenswrapper[4970]: E1124 14:36:43.472641 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:36:57 crc kubenswrapper[4970]: I1124 14:36:57.470644 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:36:57 crc kubenswrapper[4970]: E1124 14:36:57.471277 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:37:12 crc kubenswrapper[4970]: I1124 14:37:12.470744 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:37:12 crc kubenswrapper[4970]: E1124 14:37:12.471965 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:37:24 crc kubenswrapper[4970]: I1124 14:37:24.486743 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:37:24 crc kubenswrapper[4970]: E1124 14:37:24.489857 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:37:37 crc kubenswrapper[4970]: I1124 14:37:37.470962 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:37:37 crc kubenswrapper[4970]: E1124 14:37:37.473027 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:37:49 crc kubenswrapper[4970]: I1124 14:37:49.471312 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:37:49 crc kubenswrapper[4970]: E1124 14:37:49.472360 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.273264 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:37:53 crc kubenswrapper[4970]: E1124 14:37:53.274398 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="extract-utilities" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.274417 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="extract-utilities" Nov 24 14:37:53 crc kubenswrapper[4970]: E1124 14:37:53.274433 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="registry-server" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.274442 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="registry-server" Nov 24 14:37:53 crc kubenswrapper[4970]: E1124 14:37:53.274474 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="extract-content" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.274482 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="extract-content" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.274747 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="707341a3-d708-401e-8527-26947cbf3d30" containerName="registry-server" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.276519 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.297232 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.423233 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.423519 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmwgj\" (UniqueName: \"kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.423629 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.526365 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmwgj\" (UniqueName: \"kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.526756 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.527024 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.527271 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.527407 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.547271 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmwgj\" (UniqueName: \"kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj\") pod \"community-operators-x54js\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:53 crc kubenswrapper[4970]: I1124 14:37:53.605156 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:37:54 crc kubenswrapper[4970]: I1124 14:37:54.166913 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:37:54 crc kubenswrapper[4970]: I1124 14:37:54.811558 4970 generic.go:334] "Generic (PLEG): container finished" podID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerID="5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7" exitCode=0 Nov 24 14:37:54 crc kubenswrapper[4970]: I1124 14:37:54.811729 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerDied","Data":"5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7"} Nov 24 14:37:54 crc kubenswrapper[4970]: I1124 14:37:54.812036 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerStarted","Data":"f632cf6d1300b9e63bdb5b2e2bfe80de98d042da68d9de8e3419edea949bb801"} Nov 24 14:37:54 crc kubenswrapper[4970]: I1124 14:37:54.814947 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.083717 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.086632 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.096279 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.164986 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.165123 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mghtf\" (UniqueName: \"kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.165177 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.266672 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mghtf\" (UniqueName: \"kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.266757 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.266880 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.267866 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.268076 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.291879 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mghtf\" (UniqueName: \"kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf\") pod \"redhat-operators-t48nf\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.408833 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.820183 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerStarted","Data":"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c"} Nov 24 14:37:55 crc kubenswrapper[4970]: I1124 14:37:55.931184 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:37:56 crc kubenswrapper[4970]: I1124 14:37:56.828777 4970 generic.go:334] "Generic (PLEG): container finished" podID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerID="f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870" exitCode=0 Nov 24 14:37:56 crc kubenswrapper[4970]: I1124 14:37:56.828851 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerDied","Data":"f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870"} Nov 24 14:37:56 crc kubenswrapper[4970]: I1124 14:37:56.828881 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerStarted","Data":"3bdce7417cf9b0927f2018ab1e61724b8a0eda239d9c3c636b9790adb3225b0a"} Nov 24 14:37:56 crc kubenswrapper[4970]: I1124 14:37:56.832141 4970 generic.go:334] "Generic (PLEG): container finished" podID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerID="6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c" exitCode=0 Nov 24 14:37:56 crc kubenswrapper[4970]: I1124 14:37:56.832192 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerDied","Data":"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c"} Nov 24 14:37:57 crc kubenswrapper[4970]: I1124 14:37:57.844619 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerStarted","Data":"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f"} Nov 24 14:37:57 crc kubenswrapper[4970]: I1124 14:37:57.881095 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x54js" podStartSLOduration=2.458123783 podStartE2EDuration="4.881074408s" podCreationTimestamp="2025-11-24 14:37:53 +0000 UTC" firstStartedPulling="2025-11-24 14:37:54.814662505 +0000 UTC m=+3090.102419808" lastFinishedPulling="2025-11-24 14:37:57.23761314 +0000 UTC m=+3092.525370433" observedRunningTime="2025-11-24 14:37:57.873313454 +0000 UTC m=+3093.161070757" watchObservedRunningTime="2025-11-24 14:37:57.881074408 +0000 UTC m=+3093.168831701" Nov 24 14:37:58 crc kubenswrapper[4970]: I1124 14:37:58.855875 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerStarted","Data":"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb"} Nov 24 14:38:00 crc kubenswrapper[4970]: I1124 14:38:00.470751 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:38:00 crc kubenswrapper[4970]: E1124 14:38:00.471378 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:38:01 crc kubenswrapper[4970]: I1124 14:38:01.889143 4970 generic.go:334] "Generic (PLEG): container finished" podID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerID="e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb" exitCode=0 Nov 24 14:38:01 crc kubenswrapper[4970]: I1124 14:38:01.889198 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerDied","Data":"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb"} Nov 24 14:38:02 crc kubenswrapper[4970]: I1124 14:38:02.918764 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerStarted","Data":"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96"} Nov 24 14:38:02 crc kubenswrapper[4970]: I1124 14:38:02.945212 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t48nf" podStartSLOduration=2.474370172 podStartE2EDuration="7.9451921s" podCreationTimestamp="2025-11-24 14:37:55 +0000 UTC" firstStartedPulling="2025-11-24 14:37:56.831652984 +0000 UTC m=+3092.119410277" lastFinishedPulling="2025-11-24 14:38:02.302474882 +0000 UTC m=+3097.590232205" observedRunningTime="2025-11-24 14:38:02.94083374 +0000 UTC m=+3098.228591073" watchObservedRunningTime="2025-11-24 14:38:02.9451921 +0000 UTC m=+3098.232949393" Nov 24 14:38:03 crc kubenswrapper[4970]: I1124 14:38:03.606131 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:03 crc kubenswrapper[4970]: I1124 14:38:03.606519 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:03 crc kubenswrapper[4970]: I1124 14:38:03.657356 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:03 crc kubenswrapper[4970]: I1124 14:38:03.988648 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:04 crc kubenswrapper[4970]: I1124 14:38:04.868974 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:38:05 crc kubenswrapper[4970]: I1124 14:38:05.411382 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:05 crc kubenswrapper[4970]: I1124 14:38:05.411448 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:05 crc kubenswrapper[4970]: I1124 14:38:05.943034 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x54js" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="registry-server" containerID="cri-o://6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f" gracePeriod=2 Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.462599 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t48nf" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="registry-server" probeResult="failure" output=< Nov 24 14:38:06 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 14:38:06 crc kubenswrapper[4970]: > Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.484928 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.602184 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmwgj\" (UniqueName: \"kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj\") pod \"b2b80263-7313-4fd7-9e9e-4646e66827e9\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.602263 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content\") pod \"b2b80263-7313-4fd7-9e9e-4646e66827e9\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.602284 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities\") pod \"b2b80263-7313-4fd7-9e9e-4646e66827e9\" (UID: \"b2b80263-7313-4fd7-9e9e-4646e66827e9\") " Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.603628 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities" (OuterVolumeSpecName: "utilities") pod "b2b80263-7313-4fd7-9e9e-4646e66827e9" (UID: "b2b80263-7313-4fd7-9e9e-4646e66827e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.612442 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj" (OuterVolumeSpecName: "kube-api-access-mmwgj") pod "b2b80263-7313-4fd7-9e9e-4646e66827e9" (UID: "b2b80263-7313-4fd7-9e9e-4646e66827e9"). InnerVolumeSpecName "kube-api-access-mmwgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.657395 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2b80263-7313-4fd7-9e9e-4646e66827e9" (UID: "b2b80263-7313-4fd7-9e9e-4646e66827e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.706764 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmwgj\" (UniqueName: \"kubernetes.io/projected/b2b80263-7313-4fd7-9e9e-4646e66827e9-kube-api-access-mmwgj\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.707020 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.707029 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b80263-7313-4fd7-9e9e-4646e66827e9-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.952473 4970 generic.go:334] "Generic (PLEG): container finished" podID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerID="6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f" exitCode=0 Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.952525 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerDied","Data":"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f"} Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.952556 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x54js" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.952602 4970 scope.go:117] "RemoveContainer" containerID="6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.952569 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x54js" event={"ID":"b2b80263-7313-4fd7-9e9e-4646e66827e9","Type":"ContainerDied","Data":"f632cf6d1300b9e63bdb5b2e2bfe80de98d042da68d9de8e3419edea949bb801"} Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.972522 4970 scope.go:117] "RemoveContainer" containerID="6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c" Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.989077 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:38:06 crc kubenswrapper[4970]: I1124 14:38:06.996641 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x54js"] Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.004671 4970 scope.go:117] "RemoveContainer" containerID="5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.042678 4970 scope.go:117] "RemoveContainer" containerID="6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f" Nov 24 14:38:07 crc kubenswrapper[4970]: E1124 14:38:07.043140 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f\": container with ID starting with 6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f not found: ID does not exist" containerID="6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.043183 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f"} err="failed to get container status \"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f\": rpc error: code = NotFound desc = could not find container \"6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f\": container with ID starting with 6028941fe4f86e8064f0b82bfb630c6e2122f9c8aba7011532e21eec35d3357f not found: ID does not exist" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.043212 4970 scope.go:117] "RemoveContainer" containerID="6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c" Nov 24 14:38:07 crc kubenswrapper[4970]: E1124 14:38:07.043749 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c\": container with ID starting with 6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c not found: ID does not exist" containerID="6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.043800 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c"} err="failed to get container status \"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c\": rpc error: code = NotFound desc = could not find container \"6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c\": container with ID starting with 6470cfcabbfdb0edc78a1757aa88e9580dfbc3d0192e2b1481a5c6bd080d223c not found: ID does not exist" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.043833 4970 scope.go:117] "RemoveContainer" containerID="5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7" Nov 24 14:38:07 crc kubenswrapper[4970]: E1124 14:38:07.044220 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7\": container with ID starting with 5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7 not found: ID does not exist" containerID="5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.044261 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7"} err="failed to get container status \"5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7\": rpc error: code = NotFound desc = could not find container \"5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7\": container with ID starting with 5afa09bff9188b4074dce57f578367361feea4163c1b696b89378d210fc621d7 not found: ID does not exist" Nov 24 14:38:07 crc kubenswrapper[4970]: I1124 14:38:07.481645 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" path="/var/lib/kubelet/pods/b2b80263-7313-4fd7-9e9e-4646e66827e9/volumes" Nov 24 14:38:12 crc kubenswrapper[4970]: I1124 14:38:12.471255 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:38:12 crc kubenswrapper[4970]: E1124 14:38:12.472142 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:38:15 crc kubenswrapper[4970]: I1124 14:38:15.459721 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:15 crc kubenswrapper[4970]: I1124 14:38:15.517245 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:15 crc kubenswrapper[4970]: I1124 14:38:15.711777 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.054347 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t48nf" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="registry-server" containerID="cri-o://fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96" gracePeriod=2 Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.521958 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.678180 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mghtf\" (UniqueName: \"kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf\") pod \"94ff9d31-0535-4f3c-b571-e34a57272a2b\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.678418 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content\") pod \"94ff9d31-0535-4f3c-b571-e34a57272a2b\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.678497 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities\") pod \"94ff9d31-0535-4f3c-b571-e34a57272a2b\" (UID: \"94ff9d31-0535-4f3c-b571-e34a57272a2b\") " Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.679372 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities" (OuterVolumeSpecName: "utilities") pod "94ff9d31-0535-4f3c-b571-e34a57272a2b" (UID: "94ff9d31-0535-4f3c-b571-e34a57272a2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.684562 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf" (OuterVolumeSpecName: "kube-api-access-mghtf") pod "94ff9d31-0535-4f3c-b571-e34a57272a2b" (UID: "94ff9d31-0535-4f3c-b571-e34a57272a2b"). InnerVolumeSpecName "kube-api-access-mghtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.763398 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94ff9d31-0535-4f3c-b571-e34a57272a2b" (UID: "94ff9d31-0535-4f3c-b571-e34a57272a2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.780397 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.780422 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94ff9d31-0535-4f3c-b571-e34a57272a2b-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:17 crc kubenswrapper[4970]: I1124 14:38:17.780436 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mghtf\" (UniqueName: \"kubernetes.io/projected/94ff9d31-0535-4f3c-b571-e34a57272a2b-kube-api-access-mghtf\") on node \"crc\" DevicePath \"\"" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.066069 4970 generic.go:334] "Generic (PLEG): container finished" podID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerID="fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96" exitCode=0 Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.066115 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerDied","Data":"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96"} Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.066142 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t48nf" event={"ID":"94ff9d31-0535-4f3c-b571-e34a57272a2b","Type":"ContainerDied","Data":"3bdce7417cf9b0927f2018ab1e61724b8a0eda239d9c3c636b9790adb3225b0a"} Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.066161 4970 scope.go:117] "RemoveContainer" containerID="fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.066217 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t48nf" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.097041 4970 scope.go:117] "RemoveContainer" containerID="e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.108934 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.121152 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t48nf"] Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.138972 4970 scope.go:117] "RemoveContainer" containerID="f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.176422 4970 scope.go:117] "RemoveContainer" containerID="fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96" Nov 24 14:38:18 crc kubenswrapper[4970]: E1124 14:38:18.177681 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96\": container with ID starting with fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96 not found: ID does not exist" containerID="fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.177723 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96"} err="failed to get container status \"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96\": rpc error: code = NotFound desc = could not find container \"fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96\": container with ID starting with fd815c124e5b1d6d09ac1a281942234454deed4e538f0a999835011c84884a96 not found: ID does not exist" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.177749 4970 scope.go:117] "RemoveContainer" containerID="e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb" Nov 24 14:38:18 crc kubenswrapper[4970]: E1124 14:38:18.178223 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb\": container with ID starting with e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb not found: ID does not exist" containerID="e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.178283 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb"} err="failed to get container status \"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb\": rpc error: code = NotFound desc = could not find container \"e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb\": container with ID starting with e4da8dfddc8e4fa28c7e9ea038d423019acead230f118bbc841a3d15e3baf9cb not found: ID does not exist" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.178321 4970 scope.go:117] "RemoveContainer" containerID="f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870" Nov 24 14:38:18 crc kubenswrapper[4970]: E1124 14:38:18.178815 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870\": container with ID starting with f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870 not found: ID does not exist" containerID="f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870" Nov 24 14:38:18 crc kubenswrapper[4970]: I1124 14:38:18.178847 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870"} err="failed to get container status \"f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870\": rpc error: code = NotFound desc = could not find container \"f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870\": container with ID starting with f46f54eb0c8be050a6307aa1d115304ceec3f3cb14e4c199347f55c55c290870 not found: ID does not exist" Nov 24 14:38:19 crc kubenswrapper[4970]: I1124 14:38:19.483778 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" path="/var/lib/kubelet/pods/94ff9d31-0535-4f3c-b571-e34a57272a2b/volumes" Nov 24 14:38:23 crc kubenswrapper[4970]: I1124 14:38:23.470183 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:38:23 crc kubenswrapper[4970]: E1124 14:38:23.470967 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:38:38 crc kubenswrapper[4970]: I1124 14:38:38.470914 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:38:38 crc kubenswrapper[4970]: E1124 14:38:38.472114 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:38:49 crc kubenswrapper[4970]: I1124 14:38:49.470544 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:38:49 crc kubenswrapper[4970]: E1124 14:38:49.471354 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:39:04 crc kubenswrapper[4970]: I1124 14:39:04.470492 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:39:04 crc kubenswrapper[4970]: E1124 14:39:04.471400 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:39:17 crc kubenswrapper[4970]: I1124 14:39:17.470505 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:39:17 crc kubenswrapper[4970]: E1124 14:39:17.471478 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:39:32 crc kubenswrapper[4970]: I1124 14:39:32.470314 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:39:32 crc kubenswrapper[4970]: E1124 14:39:32.471253 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:39:32 crc kubenswrapper[4970]: I1124 14:39:32.827904 4970 generic.go:334] "Generic (PLEG): container finished" podID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" containerID="9aa63fc36451f266eb948fc187fa0d420c264fc9568c5b18a00ba6765605617a" exitCode=0 Nov 24 14:39:32 crc kubenswrapper[4970]: I1124 14:39:32.827962 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1ffd9b46-906d-47a2-870e-6ad3d5381c2d","Type":"ContainerDied","Data":"9aa63fc36451f266eb948fc187fa0d420c264fc9568c5b18a00ba6765605617a"} Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.277630 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451014 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451131 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451218 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451279 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451350 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451405 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451498 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmlwl\" (UniqueName: \"kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451544 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.451617 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key\") pod \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\" (UID: \"1ffd9b46-906d-47a2-870e-6ad3d5381c2d\") " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.452983 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.454156 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data" (OuterVolumeSpecName: "config-data") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.457916 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl" (OuterVolumeSpecName: "kube-api-access-cmlwl") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "kube-api-access-cmlwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.470255 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.471880 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.484794 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.487621 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.496789 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.499724 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "1ffd9b46-906d-47a2-870e-6ad3d5381c2d" (UID: "1ffd9b46-906d-47a2-870e-6ad3d5381c2d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.553526 4970 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.553815 4970 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.553878 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.553947 4970 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.554011 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmlwl\" (UniqueName: \"kubernetes.io/projected/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-kube-api-access-cmlwl\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.554070 4970 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.554132 4970 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.554184 4970 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.554240 4970 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/1ffd9b46-906d-47a2-870e-6ad3d5381c2d-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.586224 4970 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.656896 4970 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.845616 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"1ffd9b46-906d-47a2-870e-6ad3d5381c2d","Type":"ContainerDied","Data":"ffd8964ff7f85fa42f4ce7c1fa4e4456006fb7dc62f8c3d071ecf120f8550030"} Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.845941 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffd8964ff7f85fa42f4ce7c1fa4e4456006fb7dc62f8c3d071ecf120f8550030" Nov 24 14:39:34 crc kubenswrapper[4970]: I1124 14:39:34.845661 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 24 14:39:43 crc kubenswrapper[4970]: I1124 14:39:43.470744 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:39:43 crc kubenswrapper[4970]: I1124 14:39:43.938382 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8"} Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628115 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628842 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="extract-content" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628855 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="extract-content" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628887 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628895 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628909 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628916 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628930 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="extract-utilities" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628935 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="extract-utilities" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628946 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="extract-content" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628951 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="extract-content" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628969 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628976 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:39:44 crc kubenswrapper[4970]: E1124 14:39:44.628991 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="extract-utilities" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.628999 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="extract-utilities" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.629231 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b80263-7313-4fd7-9e9e-4646e66827e9" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.629261 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ff9d31-0535-4f3c-b571-e34a57272a2b" containerName="registry-server" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.629305 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ffd9b46-906d-47a2-870e-6ad3d5381c2d" containerName="tempest-tests-tempest-tests-runner" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.630085 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.632799 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2qq2r" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.667382 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.758452 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.758739 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlf82\" (UniqueName: \"kubernetes.io/projected/41506018-039f-4567-8ed7-2501e5d565a6-kube-api-access-wlf82\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.860363 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlf82\" (UniqueName: \"kubernetes.io/projected/41506018-039f-4567-8ed7-2501e5d565a6-kube-api-access-wlf82\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.860845 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.861255 4970 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.881041 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlf82\" (UniqueName: \"kubernetes.io/projected/41506018-039f-4567-8ed7-2501e5d565a6-kube-api-access-wlf82\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.892915 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"41506018-039f-4567-8ed7-2501e5d565a6\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:44 crc kubenswrapper[4970]: I1124 14:39:44.957947 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 24 14:39:45 crc kubenswrapper[4970]: I1124 14:39:45.437560 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 24 14:39:45 crc kubenswrapper[4970]: I1124 14:39:45.964939 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"41506018-039f-4567-8ed7-2501e5d565a6","Type":"ContainerStarted","Data":"5b673c6bf2f9c63eb49d9cd481bf0c10023cd3c570d77c9a3d7ffa9080caa5d0"} Nov 24 14:39:46 crc kubenswrapper[4970]: I1124 14:39:46.978074 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"41506018-039f-4567-8ed7-2501e5d565a6","Type":"ContainerStarted","Data":"8936ae3f0b52349fe3ec39d00a1631bc159b3e7d0890f6eba8d678a272be5bf3"} Nov 24 14:39:46 crc kubenswrapper[4970]: I1124 14:39:46.996215 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.8495598009999998 podStartE2EDuration="2.99618982s" podCreationTimestamp="2025-11-24 14:39:44 +0000 UTC" firstStartedPulling="2025-11-24 14:39:45.445027414 +0000 UTC m=+3200.732784747" lastFinishedPulling="2025-11-24 14:39:46.591657473 +0000 UTC m=+3201.879414766" observedRunningTime="2025-11-24 14:39:46.991770659 +0000 UTC m=+3202.279528002" watchObservedRunningTime="2025-11-24 14:39:46.99618982 +0000 UTC m=+3202.283947123" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.283455 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqf9z/must-gather-vkhbq"] Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.289774 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.292510 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wqf9z"/"openshift-service-ca.crt" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.295827 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wqf9z"/"kube-root-ca.crt" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.320419 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wqf9z/must-gather-vkhbq"] Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.455923 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.456671 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hgvs\" (UniqueName: \"kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.558520 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hgvs\" (UniqueName: \"kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.558675 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.559269 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.588125 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hgvs\" (UniqueName: \"kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs\") pod \"must-gather-vkhbq\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:09 crc kubenswrapper[4970]: I1124 14:40:09.632437 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:40:10 crc kubenswrapper[4970]: I1124 14:40:10.083395 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wqf9z/must-gather-vkhbq"] Nov 24 14:40:10 crc kubenswrapper[4970]: I1124 14:40:10.239980 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" event={"ID":"e2aa957f-38ff-46f5-9e9b-a925278d9060","Type":"ContainerStarted","Data":"a50cfc9ce556b77d7d049f83807f08f90081b0af709504eecfc6e3b779d3e5ee"} Nov 24 14:40:15 crc kubenswrapper[4970]: I1124 14:40:15.286291 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" event={"ID":"e2aa957f-38ff-46f5-9e9b-a925278d9060","Type":"ContainerStarted","Data":"d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967"} Nov 24 14:40:15 crc kubenswrapper[4970]: I1124 14:40:15.286649 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" event={"ID":"e2aa957f-38ff-46f5-9e9b-a925278d9060","Type":"ContainerStarted","Data":"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f"} Nov 24 14:40:15 crc kubenswrapper[4970]: I1124 14:40:15.304763 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" podStartSLOduration=2.270931264 podStartE2EDuration="6.304743277s" podCreationTimestamp="2025-11-24 14:40:09 +0000 UTC" firstStartedPulling="2025-11-24 14:40:10.082980019 +0000 UTC m=+3225.370737312" lastFinishedPulling="2025-11-24 14:40:14.116792012 +0000 UTC m=+3229.404549325" observedRunningTime="2025-11-24 14:40:15.303492192 +0000 UTC m=+3230.591249495" watchObservedRunningTime="2025-11-24 14:40:15.304743277 +0000 UTC m=+3230.592500580" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.036493 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-j5t7r"] Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.039257 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.043908 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wqf9z"/"default-dockercfg-vxlgn" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.057035 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdzps\" (UniqueName: \"kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.057175 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.158477 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.158620 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdzps\" (UniqueName: \"kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.158630 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.182430 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdzps\" (UniqueName: \"kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps\") pod \"crc-debug-j5t7r\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:18 crc kubenswrapper[4970]: I1124 14:40:18.355939 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:40:19 crc kubenswrapper[4970]: I1124 14:40:19.346802 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" event={"ID":"19da5311-2571-49f5-ae27-7de4e7292c0d","Type":"ContainerStarted","Data":"aa6a738ae17c7951299fba92f083dab5821a1474338524e98277907835c2bfe6"} Nov 24 14:40:29 crc kubenswrapper[4970]: I1124 14:40:29.435078 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" event={"ID":"19da5311-2571-49f5-ae27-7de4e7292c0d","Type":"ContainerStarted","Data":"16be3ae2d6669417997961cab1645ea8ac3e32d256d6b1e8ad4e531ed4740b64"} Nov 24 14:40:29 crc kubenswrapper[4970]: I1124 14:40:29.448949 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" podStartSLOduration=1.566390587 podStartE2EDuration="11.448932233s" podCreationTimestamp="2025-11-24 14:40:18 +0000 UTC" firstStartedPulling="2025-11-24 14:40:18.386414951 +0000 UTC m=+3233.674172244" lastFinishedPulling="2025-11-24 14:40:28.268956597 +0000 UTC m=+3243.556713890" observedRunningTime="2025-11-24 14:40:29.44773681 +0000 UTC m=+3244.735494113" watchObservedRunningTime="2025-11-24 14:40:29.448932233 +0000 UTC m=+3244.736689526" Nov 24 14:41:07 crc kubenswrapper[4970]: I1124 14:41:07.792715 4970 generic.go:334] "Generic (PLEG): container finished" podID="19da5311-2571-49f5-ae27-7de4e7292c0d" containerID="16be3ae2d6669417997961cab1645ea8ac3e32d256d6b1e8ad4e531ed4740b64" exitCode=0 Nov 24 14:41:07 crc kubenswrapper[4970]: I1124 14:41:07.792849 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" event={"ID":"19da5311-2571-49f5-ae27-7de4e7292c0d","Type":"ContainerDied","Data":"16be3ae2d6669417997961cab1645ea8ac3e32d256d6b1e8ad4e531ed4740b64"} Nov 24 14:41:08 crc kubenswrapper[4970]: I1124 14:41:08.943625 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:41:08 crc kubenswrapper[4970]: I1124 14:41:08.983625 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-j5t7r"] Nov 24 14:41:08 crc kubenswrapper[4970]: I1124 14:41:08.994399 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-j5t7r"] Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.025313 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host\") pod \"19da5311-2571-49f5-ae27-7de4e7292c0d\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.025392 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host" (OuterVolumeSpecName: "host") pod "19da5311-2571-49f5-ae27-7de4e7292c0d" (UID: "19da5311-2571-49f5-ae27-7de4e7292c0d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.025438 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdzps\" (UniqueName: \"kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps\") pod \"19da5311-2571-49f5-ae27-7de4e7292c0d\" (UID: \"19da5311-2571-49f5-ae27-7de4e7292c0d\") " Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.025951 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19da5311-2571-49f5-ae27-7de4e7292c0d-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.035098 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps" (OuterVolumeSpecName: "kube-api-access-sdzps") pod "19da5311-2571-49f5-ae27-7de4e7292c0d" (UID: "19da5311-2571-49f5-ae27-7de4e7292c0d"). InnerVolumeSpecName "kube-api-access-sdzps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.127786 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdzps\" (UniqueName: \"kubernetes.io/projected/19da5311-2571-49f5-ae27-7de4e7292c0d-kube-api-access-sdzps\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.490459 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19da5311-2571-49f5-ae27-7de4e7292c0d" path="/var/lib/kubelet/pods/19da5311-2571-49f5-ae27-7de4e7292c0d/volumes" Nov 24 14:41:09 crc kubenswrapper[4970]: E1124 14:41:09.656062 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19da5311_2571_49f5_ae27_7de4e7292c0d.slice\": RecentStats: unable to find data in memory cache]" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.818319 4970 scope.go:117] "RemoveContainer" containerID="16be3ae2d6669417997961cab1645ea8ac3e32d256d6b1e8ad4e531ed4740b64" Nov 24 14:41:09 crc kubenswrapper[4970]: I1124 14:41:09.818391 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-j5t7r" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.218678 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-2hwx4"] Nov 24 14:41:10 crc kubenswrapper[4970]: E1124 14:41:10.219467 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19da5311-2571-49f5-ae27-7de4e7292c0d" containerName="container-00" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.219498 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="19da5311-2571-49f5-ae27-7de4e7292c0d" containerName="container-00" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.220095 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="19da5311-2571-49f5-ae27-7de4e7292c0d" containerName="container-00" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.221252 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.223891 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wqf9z"/"default-dockercfg-vxlgn" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.372726 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.372939 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ntrs\" (UniqueName: \"kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.474716 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.474853 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ntrs\" (UniqueName: \"kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.474881 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.494105 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ntrs\" (UniqueName: \"kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs\") pod \"crc-debug-2hwx4\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.546128 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:10 crc kubenswrapper[4970]: I1124 14:41:10.835830 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" event={"ID":"d6c16eb4-13d0-420b-a983-50e9194dd679","Type":"ContainerStarted","Data":"0e8619fe5a8c0fefbe379a7b1f899d05a6f6c63c0301c0d6799b0a953ce12600"} Nov 24 14:41:11 crc kubenswrapper[4970]: I1124 14:41:11.851799 4970 generic.go:334] "Generic (PLEG): container finished" podID="d6c16eb4-13d0-420b-a983-50e9194dd679" containerID="9ebb608f7dda030585f25b1ada7eaf04688689360dfbc7a63c02ba9c74b1ab5b" exitCode=0 Nov 24 14:41:11 crc kubenswrapper[4970]: I1124 14:41:11.851865 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" event={"ID":"d6c16eb4-13d0-420b-a983-50e9194dd679","Type":"ContainerDied","Data":"9ebb608f7dda030585f25b1ada7eaf04688689360dfbc7a63c02ba9c74b1ab5b"} Nov 24 14:41:12 crc kubenswrapper[4970]: I1124 14:41:12.460828 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-2hwx4"] Nov 24 14:41:12 crc kubenswrapper[4970]: I1124 14:41:12.467847 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-2hwx4"] Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.025046 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.123897 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host\") pod \"d6c16eb4-13d0-420b-a983-50e9194dd679\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.124027 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ntrs\" (UniqueName: \"kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs\") pod \"d6c16eb4-13d0-420b-a983-50e9194dd679\" (UID: \"d6c16eb4-13d0-420b-a983-50e9194dd679\") " Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.124039 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host" (OuterVolumeSpecName: "host") pod "d6c16eb4-13d0-420b-a983-50e9194dd679" (UID: "d6c16eb4-13d0-420b-a983-50e9194dd679"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.125250 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d6c16eb4-13d0-420b-a983-50e9194dd679-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.133541 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs" (OuterVolumeSpecName: "kube-api-access-8ntrs") pod "d6c16eb4-13d0-420b-a983-50e9194dd679" (UID: "d6c16eb4-13d0-420b-a983-50e9194dd679"). InnerVolumeSpecName "kube-api-access-8ntrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.226632 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ntrs\" (UniqueName: \"kubernetes.io/projected/d6c16eb4-13d0-420b-a983-50e9194dd679-kube-api-access-8ntrs\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.481866 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6c16eb4-13d0-420b-a983-50e9194dd679" path="/var/lib/kubelet/pods/d6c16eb4-13d0-420b-a983-50e9194dd679/volumes" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.656559 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-77r69"] Nov 24 14:41:13 crc kubenswrapper[4970]: E1124 14:41:13.657112 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6c16eb4-13d0-420b-a983-50e9194dd679" containerName="container-00" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.657142 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6c16eb4-13d0-420b-a983-50e9194dd679" containerName="container-00" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.657509 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6c16eb4-13d0-420b-a983-50e9194dd679" containerName="container-00" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.658477 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.837912 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.838189 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnwtv\" (UniqueName: \"kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.878430 4970 scope.go:117] "RemoveContainer" containerID="9ebb608f7dda030585f25b1ada7eaf04688689360dfbc7a63c02ba9c74b1ab5b" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.878664 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-2hwx4" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.939924 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.940049 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.940653 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnwtv\" (UniqueName: \"kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.962019 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnwtv\" (UniqueName: \"kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv\") pod \"crc-debug-77r69\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:13 crc kubenswrapper[4970]: I1124 14:41:13.987569 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:14 crc kubenswrapper[4970]: W1124 14:41:14.025898 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c38d4b1_14eb_414f_bc67_ad36be00ba43.slice/crio-6b4fa65dc0bc0d5740a6cc37348ca486865e31229dd8b3b759a5a93a7c52547a WatchSource:0}: Error finding container 6b4fa65dc0bc0d5740a6cc37348ca486865e31229dd8b3b759a5a93a7c52547a: Status 404 returned error can't find the container with id 6b4fa65dc0bc0d5740a6cc37348ca486865e31229dd8b3b759a5a93a7c52547a Nov 24 14:41:14 crc kubenswrapper[4970]: I1124 14:41:14.893526 4970 generic.go:334] "Generic (PLEG): container finished" podID="5c38d4b1-14eb-414f-bc67-ad36be00ba43" containerID="50b7bcecf9510706a0ae23d8aa14ea3ef76ba0afa33d92b811aa8ff6b299d368" exitCode=0 Nov 24 14:41:14 crc kubenswrapper[4970]: I1124 14:41:14.893632 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-77r69" event={"ID":"5c38d4b1-14eb-414f-bc67-ad36be00ba43","Type":"ContainerDied","Data":"50b7bcecf9510706a0ae23d8aa14ea3ef76ba0afa33d92b811aa8ff6b299d368"} Nov 24 14:41:14 crc kubenswrapper[4970]: I1124 14:41:14.894076 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/crc-debug-77r69" event={"ID":"5c38d4b1-14eb-414f-bc67-ad36be00ba43","Type":"ContainerStarted","Data":"6b4fa65dc0bc0d5740a6cc37348ca486865e31229dd8b3b759a5a93a7c52547a"} Nov 24 14:41:14 crc kubenswrapper[4970]: I1124 14:41:14.959529 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-77r69"] Nov 24 14:41:14 crc kubenswrapper[4970]: I1124 14:41:14.973986 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqf9z/crc-debug-77r69"] Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.014456 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.186017 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnwtv\" (UniqueName: \"kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv\") pod \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.186221 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host\") pod \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\" (UID: \"5c38d4b1-14eb-414f-bc67-ad36be00ba43\") " Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.186515 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host" (OuterVolumeSpecName: "host") pod "5c38d4b1-14eb-414f-bc67-ad36be00ba43" (UID: "5c38d4b1-14eb-414f-bc67-ad36be00ba43"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.192605 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv" (OuterVolumeSpecName: "kube-api-access-jnwtv") pod "5c38d4b1-14eb-414f-bc67-ad36be00ba43" (UID: "5c38d4b1-14eb-414f-bc67-ad36be00ba43"). InnerVolumeSpecName "kube-api-access-jnwtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.288305 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnwtv\" (UniqueName: \"kubernetes.io/projected/5c38d4b1-14eb-414f-bc67-ad36be00ba43-kube-api-access-jnwtv\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.288354 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5c38d4b1-14eb-414f-bc67-ad36be00ba43-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.918954 4970 scope.go:117] "RemoveContainer" containerID="50b7bcecf9510706a0ae23d8aa14ea3ef76ba0afa33d92b811aa8ff6b299d368" Nov 24 14:41:16 crc kubenswrapper[4970]: I1124 14:41:16.919011 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/crc-debug-77r69" Nov 24 14:41:17 crc kubenswrapper[4970]: I1124 14:41:17.480867 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c38d4b1-14eb-414f-bc67-ad36be00ba43" path="/var/lib/kubelet/pods/5c38d4b1-14eb-414f-bc67-ad36be00ba43/volumes" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.384632 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756b7f5776-9gksb_83e2deca-2802-4021-9e47-b933ace28f4a/barbican-api/0.log" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.645400 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756b7f5776-9gksb_83e2deca-2802-4021-9e47-b933ace28f4a/barbican-api-log/0.log" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.759564 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-688cdc6cb6-fkrjb_df1d70bf-d612-4ef1-a113-6be419867e1d/barbican-keystone-listener/0.log" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.773473 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-688cdc6cb6-fkrjb_df1d70bf-d612-4ef1-a113-6be419867e1d/barbican-keystone-listener-log/0.log" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.939472 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-85596cc87-lvc2b_17b7d151-f71d-4f02-9d52-4c7ba81ab89e/barbican-worker/0.log" Nov 24 14:41:31 crc kubenswrapper[4970]: I1124 14:41:31.978173 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-85596cc87-lvc2b_17b7d151-f71d-4f02-9d52-4c7ba81ab89e/barbican-worker-log/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.105528 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv_b9700095-d8fb-496e-981d-de1e30ad7e8a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.160478 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/ceilometer-central-agent/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.233404 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/ceilometer-notification-agent/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.298880 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/proxy-httpd/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.353842 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/sg-core/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.449808 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9c6c4a13-7335-4e6f-b424-90a6a003a894/cinder-api/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.556131 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9c6c4a13-7335-4e6f-b424-90a6a003a894/cinder-api-log/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.661163 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c8304c7a-67d5-4d0f-afce-231f09bf48ea/cinder-scheduler/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.691106 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c8304c7a-67d5-4d0f-afce-231f09bf48ea/probe/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.787952 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-s49zl_34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.952897 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z_35b1536c-4ac4-42f5-9f1f-09838c75d1b7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:32 crc kubenswrapper[4970]: I1124 14:41:32.984855 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/init/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.227510 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/dnsmasq-dns/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.277227 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/init/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.285496 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8_bdec1f38-8d93-4298-ad14-872b58cff4e9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.461782 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0713092e-39d2-40e1-b708-a2e2addd8249/glance-httpd/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.503341 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0713092e-39d2-40e1-b708-a2e2addd8249/glance-log/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.662219 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_639b382e-fe48-4188-ba6b-1b2c5688898c/glance-httpd/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.679560 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_639b382e-fe48-4188-ba6b-1b2c5688898c/glance-log/0.log" Nov 24 14:41:33 crc kubenswrapper[4970]: I1124 14:41:33.814334 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bb8f9fccd-2m5xk_070dfeee-2ee7-4b9e-bdbf-7320a48a0659/horizon/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.020652 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p_a87a7cba-b0be-443b-8393-6f0f21e1ccb9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.057348 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bb8f9fccd-2m5xk_070dfeee-2ee7-4b9e-bdbf-7320a48a0659/horizon-log/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.242191 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fctzs_50a618ec-6a8b-4234-8000-253f692b204a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.414691 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8c975fcc4-8cmzb_24ee1a5e-62c4-42f5-999b-27ddf44b33ae/keystone-api/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.469160 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3753fd35-b151-4bde-aa8c-a9c5f0a65f3f/kube-state-metrics/0.log" Nov 24 14:41:34 crc kubenswrapper[4970]: I1124 14:41:34.862808 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv_5dcb0227-fbb2-492f-819b-02dd5de1a728/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.086990 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6464bbf59c-l44tb_ca5d8b57-826f-4811-a8b4-4f48e3082473/neutron-httpd/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.246229 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6464bbf59c-l44tb_ca5d8b57-826f-4811-a8b4-4f48e3082473/neutron-api/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.285421 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw_c1a37385-6027-4ee8-ad98-269f93d2c46d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.775405 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_05e7b6fd-92df-4c42-b84e-151948459084/nova-api-log/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.784155 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e8645891-3d8f-4506-8139-d13f7ee73054/nova-cell0-conductor-conductor/0.log" Nov 24 14:41:35 crc kubenswrapper[4970]: I1124 14:41:35.926860 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_05e7b6fd-92df-4c42-b84e-151948459084/nova-api-api/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.033279 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0c35af6c-0331-4c98-9cb4-116c2ed45ec7/nova-cell1-conductor-conductor/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.084816 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5271e19c-5bd9-4c25-a6b9-a919f1445cca/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.293607 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-xmx7z_cfcd6f60-0b2f-4066-bb72-9634b21eb168/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.417428 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9244354d-e3fc-411a-bdc4-2d81cc2b72a0/nova-metadata-log/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.684945 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_92b5c70e-7c5c-465f-a44c-adc0aed6c664/nova-scheduler-scheduler/0.log" Nov 24 14:41:36 crc kubenswrapper[4970]: I1124 14:41:36.816330 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/mysql-bootstrap/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.009323 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/mysql-bootstrap/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.032225 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/galera/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.200231 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/mysql-bootstrap/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.390967 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9244354d-e3fc-411a-bdc4-2d81cc2b72a0/nova-metadata-metadata/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.503220 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/galera/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.510953 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/mysql-bootstrap/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.641153 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0c4f17a4-9a8d-4824-8370-6aea04d166ae/openstackclient/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.743229 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-78b67_bb1eaff5-0428-4ab0-ae3b-21f192c95279/ovn-controller/0.log" Nov 24 14:41:37 crc kubenswrapper[4970]: I1124 14:41:37.859816 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vz6qv_3f1018f7-87e9-49d7-aa02-ee711b839d82/openstack-network-exporter/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.132099 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server-init/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.246203 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.267692 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server-init/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.339560 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovs-vswitchd/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.464401 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7dr7t_1843fb9f-02d8-4021-8ef5-f5416f439ef5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.538929 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8561d773-9da8-4031-92c2-84ebd42a1afc/openstack-network-exporter/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.647016 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8561d773-9da8-4031-92c2-84ebd42a1afc/ovn-northd/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.763405 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90130f89-b1ad-43b9-9ab4-296530171bd6/openstack-network-exporter/0.log" Nov 24 14:41:38 crc kubenswrapper[4970]: I1124 14:41:38.798484 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90130f89-b1ad-43b9-9ab4-296530171bd6/ovsdbserver-nb/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.010912 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4910fec1-46d5-419d-baf9-3fa44a49b1d5/openstack-network-exporter/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.051070 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4910fec1-46d5-419d-baf9-3fa44a49b1d5/ovsdbserver-sb/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.270300 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/setup-container/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.322277 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd4546cfb-h57bs_12d2adeb-5c77-483a-bb89-63b7a50c4935/placement-api/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.324771 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd4546cfb-h57bs_12d2adeb-5c77-483a-bb89-63b7a50c4935/placement-log/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.524470 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/rabbitmq/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.587553 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/setup-container/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.617996 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/setup-container/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.798781 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/rabbitmq/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.804773 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/setup-container/0.log" Nov 24 14:41:39 crc kubenswrapper[4970]: I1124 14:41:39.841637 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m_ee264c1e-7f6a-4dc1-b406-aced8f917394/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.289157 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-c9km7_f6e35ba6-d32d-4204-a731-7f0e24802382/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.305248 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r_dafb717b-e6d5-41f4-8dab-acbd1db9f563/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.469022 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zb64x_03e03094-db8d-4000-84ea-23ace55b1768/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.528554 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pxxvx_6ac1b7b7-0e52-40cc-bd5e-042430621073/ssh-known-hosts-edpm-deployment/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.745482 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76bd896ccf-szjxt_32811b6b-86a2-4762-94fd-2eef9eedad76/proxy-server/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.768517 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76bd896ccf-szjxt_32811b6b-86a2-4762-94fd-2eef9eedad76/proxy-httpd/0.log" Nov 24 14:41:40 crc kubenswrapper[4970]: I1124 14:41:40.857890 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t72v9_b596f962-fb31-4389-937e-f42750aeaf6a/swift-ring-rebalance/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.005560 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-auditor/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.016324 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-reaper/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.086659 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-replicator/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.197004 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-auditor/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.218040 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-replicator/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.229275 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-server/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.299023 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-server/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.401659 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-updater/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.451977 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-expirer/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.486798 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-auditor/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.582443 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-replicator/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.600380 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-server/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.697015 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-updater/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.752526 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/rsync/0.log" Nov 24 14:41:41 crc kubenswrapper[4970]: I1124 14:41:41.963297 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/swift-recon-cron/0.log" Nov 24 14:41:42 crc kubenswrapper[4970]: I1124 14:41:42.087216 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-whzwn_196cf6ec-261f-4c0e-87f4-30c0e60e7bd2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:42 crc kubenswrapper[4970]: I1124 14:41:42.229404 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1ffd9b46-906d-47a2-870e-6ad3d5381c2d/tempest-tests-tempest-tests-runner/0.log" Nov 24 14:41:42 crc kubenswrapper[4970]: I1124 14:41:42.314903 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_41506018-039f-4567-8ed7-2501e5d565a6/test-operator-logs-container/0.log" Nov 24 14:41:42 crc kubenswrapper[4970]: I1124 14:41:42.499316 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w_4665b179-69e1-41a9-a621-5681262a765a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:41:49 crc kubenswrapper[4970]: I1124 14:41:49.155819 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_40d13307-a86f-453f-a3c8-05f936d03e73/memcached/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.455845 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.650216 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.697901 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.716252 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.832795 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.871677 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:42:06 crc kubenswrapper[4970]: I1124 14:42:06.872697 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/extract/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.011739 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kdx5m_93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.094070 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kdx5m_93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.123235 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-r452c_8874383f-46c3-43b7-b1d1-56c9d14a0004/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.221989 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-r452c_8874383f-46c3-43b7-b1d1-56c9d14a0004/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.277015 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-ff65l_799af0bf-24f0-4546-960b-2e8e72083fb8/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.296122 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-ff65l_799af0bf-24f0-4546-960b-2e8e72083fb8/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.421863 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qclm2_a5f4428d-ecd0-4c71-b9da-de188497a191/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.542103 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qclm2_a5f4428d-ecd0-4c71-b9da-de188497a191/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.605951 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-v67fx_9ca1ba14-aafe-429b-be3d-87dadeb73b54/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.646768 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-v67fx_9ca1ba14-aafe-429b-be3d-87dadeb73b54/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.705603 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-crqqj_4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd/kube-rbac-proxy/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.833323 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-crqqj_4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd/manager/0.log" Nov 24 14:42:07 crc kubenswrapper[4970]: I1124 14:42:07.871411 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-9lhqx_d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.056985 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-9lhqx_d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.069234 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-sx2v5_f01cc05b-92d6-4144-ad87-5fdec63cc95c/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.072261 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-sx2v5_f01cc05b-92d6-4144-ad87-5fdec63cc95c/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.205752 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-9mnkp_20a43498-16e1-46d8-9487-b87ff9cad8f8/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.321315 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-9mnkp_20a43498-16e1-46d8-9487-b87ff9cad8f8/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.370813 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-h9j24_73f36b2f-3ac8-4d5c-859b-1e376b9a5194/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.380723 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-h9j24_73f36b2f-3ac8-4d5c-859b-1e376b9a5194/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.501828 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-wfsc5_b070979e-cf9f-4793-a50a-f0e67a48a58f/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.560958 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-wfsc5_b070979e-cf9f-4793-a50a-f0e67a48a58f/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.617001 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-zkcwk_028f94a9-1db9-46e8-93f1-106d959ebe9b/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.742828 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-zkcwk_028f94a9-1db9-46e8-93f1-106d959ebe9b/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.757842 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-vs899_a44bf58b-6eb3-4bdc-ad1c-41aa7623666b/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.848993 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-vs899_a44bf58b-6eb3-4bdc-ad1c-41aa7623666b/manager/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.929614 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-g8kmd_d94c9e13-b160-4f65-a97c-9bfdd4799eb0/kube-rbac-proxy/0.log" Nov 24 14:42:08 crc kubenswrapper[4970]: I1124 14:42:08.935855 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-g8kmd_d94c9e13-b160-4f65-a97c-9bfdd4799eb0/manager/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.087255 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp_d17a257a-f687-4d53-9b81-5fa430e24d70/kube-rbac-proxy/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.102609 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp_d17a257a-f687-4d53-9b81-5fa430e24d70/manager/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.302993 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-bcbxx_6ed5751d-33ce-4775-a0c7-cb951a6fbdcb/kube-rbac-proxy/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.488788 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-pb4vh_f268a9ff-bab7-41a4-9d10-d0494d511f3a/kube-rbac-proxy/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.655962 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-pb4vh_f268a9ff-bab7-41a4-9d10-d0494d511f3a/operator/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.715550 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wrstd_2bca4967-78c0-444a-bd32-e12f72f64629/registry-server/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.897992 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-hdflg_1af43cf6-5835-47eb-a315-c16bf5758a9e/kube-rbac-proxy/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.953777 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-t4pqg_90e1030d-d9d0-42bf-b35b-a2f19ee0e6de/kube-rbac-proxy/0.log" Nov 24 14:42:09 crc kubenswrapper[4970]: I1124 14:42:09.961127 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-hdflg_1af43cf6-5835-47eb-a315-c16bf5758a9e/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.099720 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-t4pqg_90e1030d-d9d0-42bf-b35b-a2f19ee0e6de/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.189519 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-488cd_65008609-c137-4dee-b406-159814902ff4/operator/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.294317 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-5sp97_0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20/kube-rbac-proxy/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.309725 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-bcbxx_6ed5751d-33ce-4775-a0c7-cb951a6fbdcb/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.437163 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-5sp97_0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.477402 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-k5b5j_289cc484-057a-4107-937d-6c532c21b52a/kube-rbac-proxy/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.502124 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-k5b5j_289cc484-057a-4107-937d-6c532c21b52a/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.624904 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-wfmzk_5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7/kube-rbac-proxy/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.631900 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-wfmzk_5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7/manager/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.692808 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-m9m2l_9847f51b-ce73-493a-ac47-b93341cebf73/kube-rbac-proxy/0.log" Nov 24 14:42:10 crc kubenswrapper[4970]: I1124 14:42:10.717516 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-m9m2l_9847f51b-ce73-493a-ac47-b93341cebf73/manager/0.log" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.203750 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.203807 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.422407 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:11 crc kubenswrapper[4970]: E1124 14:42:11.422816 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c38d4b1-14eb-414f-bc67-ad36be00ba43" containerName="container-00" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.422832 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c38d4b1-14eb-414f-bc67-ad36be00ba43" containerName="container-00" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.423003 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c38d4b1-14eb-414f-bc67-ad36be00ba43" containerName="container-00" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.424247 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.436116 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.548675 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.549013 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.549073 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptlhj\" (UniqueName: \"kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.650553 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptlhj\" (UniqueName: \"kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.650828 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.650925 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.651650 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.651705 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.681705 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptlhj\" (UniqueName: \"kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj\") pod \"redhat-marketplace-5zv45\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:11 crc kubenswrapper[4970]: I1124 14:42:11.750311 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:12 crc kubenswrapper[4970]: I1124 14:42:12.231009 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:12 crc kubenswrapper[4970]: I1124 14:42:12.444921 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerStarted","Data":"f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3"} Nov 24 14:42:13 crc kubenswrapper[4970]: I1124 14:42:13.455265 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ba20f60-4645-4688-8665-d238986e0b8d" containerID="03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3" exitCode=0 Nov 24 14:42:13 crc kubenswrapper[4970]: I1124 14:42:13.455326 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerDied","Data":"03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3"} Nov 24 14:42:14 crc kubenswrapper[4970]: I1124 14:42:14.465748 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ba20f60-4645-4688-8665-d238986e0b8d" containerID="fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2" exitCode=0 Nov 24 14:42:14 crc kubenswrapper[4970]: I1124 14:42:14.465787 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerDied","Data":"fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2"} Nov 24 14:42:15 crc kubenswrapper[4970]: I1124 14:42:15.481105 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerStarted","Data":"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129"} Nov 24 14:42:15 crc kubenswrapper[4970]: I1124 14:42:15.500143 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5zv45" podStartSLOduration=3.097354483 podStartE2EDuration="4.500124554s" podCreationTimestamp="2025-11-24 14:42:11 +0000 UTC" firstStartedPulling="2025-11-24 14:42:13.45745924 +0000 UTC m=+3348.745216533" lastFinishedPulling="2025-11-24 14:42:14.860229311 +0000 UTC m=+3350.147986604" observedRunningTime="2025-11-24 14:42:15.498959641 +0000 UTC m=+3350.786716934" watchObservedRunningTime="2025-11-24 14:42:15.500124554 +0000 UTC m=+3350.787881847" Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.804066 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.806104 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.833231 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.914843 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.914940 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:17 crc kubenswrapper[4970]: I1124 14:42:17.914976 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rllbb\" (UniqueName: \"kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.017087 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.017183 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.017216 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rllbb\" (UniqueName: \"kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.017690 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.017682 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.047786 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rllbb\" (UniqueName: \"kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb\") pod \"certified-operators-2cvq5\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.128628 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:18 crc kubenswrapper[4970]: I1124 14:42:18.682466 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:19 crc kubenswrapper[4970]: I1124 14:42:19.531317 4970 generic.go:334] "Generic (PLEG): container finished" podID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerID="fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863" exitCode=0 Nov 24 14:42:19 crc kubenswrapper[4970]: I1124 14:42:19.531399 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerDied","Data":"fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863"} Nov 24 14:42:19 crc kubenswrapper[4970]: I1124 14:42:19.531462 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerStarted","Data":"02be358984e601786e1f7404ce7b2c1b4c5995fa4687f28b05a9ba190235f3f4"} Nov 24 14:42:20 crc kubenswrapper[4970]: I1124 14:42:20.565550 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerStarted","Data":"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632"} Nov 24 14:42:21 crc kubenswrapper[4970]: I1124 14:42:21.582501 4970 generic.go:334] "Generic (PLEG): container finished" podID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerID="20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632" exitCode=0 Nov 24 14:42:21 crc kubenswrapper[4970]: I1124 14:42:21.582655 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerDied","Data":"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632"} Nov 24 14:42:21 crc kubenswrapper[4970]: I1124 14:42:21.751317 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:21 crc kubenswrapper[4970]: I1124 14:42:21.752187 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:21 crc kubenswrapper[4970]: I1124 14:42:21.819855 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:22 crc kubenswrapper[4970]: I1124 14:42:22.597466 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerStarted","Data":"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c"} Nov 24 14:42:22 crc kubenswrapper[4970]: I1124 14:42:22.625895 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2cvq5" podStartSLOduration=2.9047015 podStartE2EDuration="5.625874386s" podCreationTimestamp="2025-11-24 14:42:17 +0000 UTC" firstStartedPulling="2025-11-24 14:42:19.535186173 +0000 UTC m=+3354.822943496" lastFinishedPulling="2025-11-24 14:42:22.256359049 +0000 UTC m=+3357.544116382" observedRunningTime="2025-11-24 14:42:22.616931061 +0000 UTC m=+3357.904688374" watchObservedRunningTime="2025-11-24 14:42:22.625874386 +0000 UTC m=+3357.913631679" Nov 24 14:42:22 crc kubenswrapper[4970]: I1124 14:42:22.671088 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:24 crc kubenswrapper[4970]: I1124 14:42:24.207490 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:24 crc kubenswrapper[4970]: I1124 14:42:24.622183 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5zv45" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="registry-server" containerID="cri-o://a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129" gracePeriod=2 Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.099290 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.177764 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptlhj\" (UniqueName: \"kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj\") pod \"5ba20f60-4645-4688-8665-d238986e0b8d\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.177943 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities\") pod \"5ba20f60-4645-4688-8665-d238986e0b8d\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.178020 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content\") pod \"5ba20f60-4645-4688-8665-d238986e0b8d\" (UID: \"5ba20f60-4645-4688-8665-d238986e0b8d\") " Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.178995 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities" (OuterVolumeSpecName: "utilities") pod "5ba20f60-4645-4688-8665-d238986e0b8d" (UID: "5ba20f60-4645-4688-8665-d238986e0b8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.202301 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj" (OuterVolumeSpecName: "kube-api-access-ptlhj") pod "5ba20f60-4645-4688-8665-d238986e0b8d" (UID: "5ba20f60-4645-4688-8665-d238986e0b8d"). InnerVolumeSpecName "kube-api-access-ptlhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.209677 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ba20f60-4645-4688-8665-d238986e0b8d" (UID: "5ba20f60-4645-4688-8665-d238986e0b8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.280043 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.280087 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ba20f60-4645-4688-8665-d238986e0b8d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.280101 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptlhj\" (UniqueName: \"kubernetes.io/projected/5ba20f60-4645-4688-8665-d238986e0b8d-kube-api-access-ptlhj\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.633555 4970 generic.go:334] "Generic (PLEG): container finished" podID="5ba20f60-4645-4688-8665-d238986e0b8d" containerID="a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129" exitCode=0 Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.633607 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerDied","Data":"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129"} Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.634203 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5zv45" event={"ID":"5ba20f60-4645-4688-8665-d238986e0b8d","Type":"ContainerDied","Data":"f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3"} Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.633670 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5zv45" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.634226 4970 scope.go:117] "RemoveContainer" containerID="a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.654786 4970 scope.go:117] "RemoveContainer" containerID="fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.659905 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.669553 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5zv45"] Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.681531 4970 scope.go:117] "RemoveContainer" containerID="03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.728513 4970 scope.go:117] "RemoveContainer" containerID="a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129" Nov 24 14:42:25 crc kubenswrapper[4970]: E1124 14:42:25.729100 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129\": container with ID starting with a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129 not found: ID does not exist" containerID="a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.729146 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129"} err="failed to get container status \"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129\": rpc error: code = NotFound desc = could not find container \"a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129\": container with ID starting with a0e97277cc6839239b8c0a1890791fd3e819c2b3e3187d8186ba8eef75d29129 not found: ID does not exist" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.729182 4970 scope.go:117] "RemoveContainer" containerID="fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2" Nov 24 14:42:25 crc kubenswrapper[4970]: E1124 14:42:25.729624 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2\": container with ID starting with fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2 not found: ID does not exist" containerID="fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.729677 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2"} err="failed to get container status \"fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2\": rpc error: code = NotFound desc = could not find container \"fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2\": container with ID starting with fa9b3a03f21bb10759333f7a297546790682ade9a6c2798914733961215322a2 not found: ID does not exist" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.729720 4970 scope.go:117] "RemoveContainer" containerID="03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3" Nov 24 14:42:25 crc kubenswrapper[4970]: E1124 14:42:25.729980 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3\": container with ID starting with 03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3 not found: ID does not exist" containerID="03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3" Nov 24 14:42:25 crc kubenswrapper[4970]: I1124 14:42:25.730055 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3"} err="failed to get container status \"03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3\": rpc error: code = NotFound desc = could not find container \"03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3\": container with ID starting with 03f97d09498d8d86342001093d2943dcdf720317f62a8668db81de52ca76e0b3 not found: ID does not exist" Nov 24 14:42:27 crc kubenswrapper[4970]: I1124 14:42:27.480168 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" path="/var/lib/kubelet/pods/5ba20f60-4645-4688-8665-d238986e0b8d/volumes" Nov 24 14:42:27 crc kubenswrapper[4970]: I1124 14:42:27.515835 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xl78j_4b2d5731-75aa-457d-8afb-75815fd5ca93/control-plane-machine-set-operator/0.log" Nov 24 14:42:27 crc kubenswrapper[4970]: I1124 14:42:27.679689 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5jd8l_3999597f-f349-4448-b4f4-45fa925f3342/machine-api-operator/0.log" Nov 24 14:42:27 crc kubenswrapper[4970]: I1124 14:42:27.686226 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5jd8l_3999597f-f349-4448-b4f4-45fa925f3342/kube-rbac-proxy/0.log" Nov 24 14:42:28 crc kubenswrapper[4970]: I1124 14:42:28.130751 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:28 crc kubenswrapper[4970]: I1124 14:42:28.130798 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:28 crc kubenswrapper[4970]: I1124 14:42:28.184458 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:28 crc kubenswrapper[4970]: I1124 14:42:28.718128 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:29 crc kubenswrapper[4970]: I1124 14:42:29.208658 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:30 crc kubenswrapper[4970]: I1124 14:42:30.677341 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2cvq5" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="registry-server" containerID="cri-o://d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c" gracePeriod=2 Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.207400 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.289120 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rllbb\" (UniqueName: \"kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb\") pod \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.289213 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities\") pod \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.289273 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content\") pod \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\" (UID: \"b3e2ce47-7b41-4e99-b2de-90b261ce0c21\") " Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.290137 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities" (OuterVolumeSpecName: "utilities") pod "b3e2ce47-7b41-4e99-b2de-90b261ce0c21" (UID: "b3e2ce47-7b41-4e99-b2de-90b261ce0c21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.295027 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb" (OuterVolumeSpecName: "kube-api-access-rllbb") pod "b3e2ce47-7b41-4e99-b2de-90b261ce0c21" (UID: "b3e2ce47-7b41-4e99-b2de-90b261ce0c21"). InnerVolumeSpecName "kube-api-access-rllbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.335191 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3e2ce47-7b41-4e99-b2de-90b261ce0c21" (UID: "b3e2ce47-7b41-4e99-b2de-90b261ce0c21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.391631 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rllbb\" (UniqueName: \"kubernetes.io/projected/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-kube-api-access-rllbb\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.391662 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.391674 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3e2ce47-7b41-4e99-b2de-90b261ce0c21-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.690699 4970 generic.go:334] "Generic (PLEG): container finished" podID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerID="d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c" exitCode=0 Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.690764 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerDied","Data":"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c"} Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.690806 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2cvq5" event={"ID":"b3e2ce47-7b41-4e99-b2de-90b261ce0c21","Type":"ContainerDied","Data":"02be358984e601786e1f7404ce7b2c1b4c5995fa4687f28b05a9ba190235f3f4"} Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.690836 4970 scope.go:117] "RemoveContainer" containerID="d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.691022 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2cvq5" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.724699 4970 scope.go:117] "RemoveContainer" containerID="20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.740230 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.749476 4970 scope.go:117] "RemoveContainer" containerID="fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.761827 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2cvq5"] Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.809180 4970 scope.go:117] "RemoveContainer" containerID="d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c" Nov 24 14:42:31 crc kubenswrapper[4970]: E1124 14:42:31.810494 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c\": container with ID starting with d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c not found: ID does not exist" containerID="d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.810542 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c"} err="failed to get container status \"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c\": rpc error: code = NotFound desc = could not find container \"d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c\": container with ID starting with d6e618f50e9a15b4f610508da4cd6a1a565fd505287b4665f0184e57e9ba056c not found: ID does not exist" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.810598 4970 scope.go:117] "RemoveContainer" containerID="20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632" Nov 24 14:42:31 crc kubenswrapper[4970]: E1124 14:42:31.810868 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632\": container with ID starting with 20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632 not found: ID does not exist" containerID="20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.810895 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632"} err="failed to get container status \"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632\": rpc error: code = NotFound desc = could not find container \"20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632\": container with ID starting with 20609e834cdc3070f6c11c9cc8430db28cf22ef54b5671201f430a008028f632 not found: ID does not exist" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.810910 4970 scope.go:117] "RemoveContainer" containerID="fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863" Nov 24 14:42:31 crc kubenswrapper[4970]: E1124 14:42:31.811127 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863\": container with ID starting with fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863 not found: ID does not exist" containerID="fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863" Nov 24 14:42:31 crc kubenswrapper[4970]: I1124 14:42:31.811153 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863"} err="failed to get container status \"fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863\": rpc error: code = NotFound desc = could not find container \"fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863\": container with ID starting with fa5c9b24d45f564030aa468447b884de8d3d644ca2af8e76c0ffa3b9d56ad863 not found: ID does not exist" Nov 24 14:42:31 crc kubenswrapper[4970]: E1124 14:42:31.836478 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache]" Nov 24 14:42:33 crc kubenswrapper[4970]: I1124 14:42:33.482406 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" path="/var/lib/kubelet/pods/b3e2ce47-7b41-4e99-b2de-90b261ce0c21/volumes" Nov 24 14:42:40 crc kubenswrapper[4970]: I1124 14:42:40.260516 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-cjd48_45373789-c5a3-47eb-b2fe-7b4ce5236a33/cert-manager-controller/0.log" Nov 24 14:42:40 crc kubenswrapper[4970]: I1124 14:42:40.420217 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-22n4h_5da63529-7e53-4ad3-8d4b-0bc32d616af8/cert-manager-cainjector/0.log" Nov 24 14:42:40 crc kubenswrapper[4970]: I1124 14:42:40.443669 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-bz6q5_fe636fdb-0614-4cbc-a621-086b3c591926/cert-manager-webhook/0.log" Nov 24 14:42:41 crc kubenswrapper[4970]: I1124 14:42:41.204422 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:42:41 crc kubenswrapper[4970]: I1124 14:42:41.204487 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:42:42 crc kubenswrapper[4970]: E1124 14:42:42.121380 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache]" Nov 24 14:42:52 crc kubenswrapper[4970]: E1124 14:42:52.404446 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache]" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.286892 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-nwkjr_ebda0f09-edf4-40b6-a13d-f04cf3ddd510/nmstate-console-plugin/0.log" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.388733 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fpsdv_2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48/nmstate-handler/0.log" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.441429 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8nmq4_62a1e00f-7e1a-4e97-a311-5628f78a7798/kube-rbac-proxy/0.log" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.543690 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8nmq4_62a1e00f-7e1a-4e97-a311-5628f78a7798/nmstate-metrics/0.log" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.620325 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-fdjzs_63629bdf-0eb6-46d8-a10b-08cc587486a1/nmstate-operator/0.log" Nov 24 14:42:53 crc kubenswrapper[4970]: I1124 14:42:53.733892 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-bjrv5_892ed300-f0bb-49aa-80e3-b7feb492606c/nmstate-webhook/0.log" Nov 24 14:43:02 crc kubenswrapper[4970]: E1124 14:43:02.765898 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache]" Nov 24 14:43:07 crc kubenswrapper[4970]: I1124 14:43:07.925849 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-s8fl4_48658fb0-3b0d-4505-9e99-68d21fa216e8/kube-rbac-proxy/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.092994 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-s8fl4_48658fb0-3b0d-4505-9e99-68d21fa216e8/controller/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.114820 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.275589 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.286509 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.310699 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.357451 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.478359 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.478359 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.507238 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.538262 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.700478 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.720025 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.722448 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/controller/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.726912 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.879618 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/frr-metrics/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.898826 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/kube-rbac-proxy/0.log" Nov 24 14:43:08 crc kubenswrapper[4970]: I1124 14:43:08.979287 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/kube-rbac-proxy-frr/0.log" Nov 24 14:43:09 crc kubenswrapper[4970]: I1124 14:43:09.088689 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/reloader/0.log" Nov 24 14:43:09 crc kubenswrapper[4970]: I1124 14:43:09.224030 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-c767v_fdf01015-8aee-4c8d-bdd2-0a72e6387d32/frr-k8s-webhook-server/0.log" Nov 24 14:43:09 crc kubenswrapper[4970]: I1124 14:43:09.465807 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56999fcf9c-k9nw7_39678371-0ddb-40cb-8b7d-375d70868b75/manager/0.log" Nov 24 14:43:09 crc kubenswrapper[4970]: I1124 14:43:09.515051 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56df5758cc-gjrmn_db2d684a-3178-4375-b857-fc4ab433a328/webhook-server/0.log" Nov 24 14:43:09 crc kubenswrapper[4970]: I1124 14:43:09.732196 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qn5c2_2bc53266-821b-4f2b-af20-c290fff7e4a9/kube-rbac-proxy/0.log" Nov 24 14:43:10 crc kubenswrapper[4970]: I1124 14:43:10.124509 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qn5c2_2bc53266-821b-4f2b-af20-c290fff7e4a9/speaker/0.log" Nov 24 14:43:10 crc kubenswrapper[4970]: I1124 14:43:10.158906 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/frr/0.log" Nov 24 14:43:11 crc kubenswrapper[4970]: I1124 14:43:11.203891 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:43:11 crc kubenswrapper[4970]: I1124 14:43:11.204205 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:43:11 crc kubenswrapper[4970]: I1124 14:43:11.204262 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:43:11 crc kubenswrapper[4970]: I1124 14:43:11.205350 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:43:11 crc kubenswrapper[4970]: I1124 14:43:11.205424 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8" gracePeriod=600 Nov 24 14:43:12 crc kubenswrapper[4970]: I1124 14:43:12.082713 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8" exitCode=0 Nov 24 14:43:12 crc kubenswrapper[4970]: I1124 14:43:12.082825 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8"} Nov 24 14:43:12 crc kubenswrapper[4970]: I1124 14:43:12.083182 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143"} Nov 24 14:43:12 crc kubenswrapper[4970]: I1124 14:43:12.083212 4970 scope.go:117] "RemoveContainer" containerID="22227d2095becec38dce7a12533ce2393a5dd0b57893eb1ebf65dc496be5e709" Nov 24 14:43:13 crc kubenswrapper[4970]: E1124 14:43:13.043902 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache]" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.011615 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.224316 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.263146 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.290945 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: E1124 14:43:23.311540 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ba20f60_4645_4688_8665_d238986e0b8d.slice/crio-f1e2fd307edef1c808810187e4c1a017b0d0c21490aaeb1ecf5b904788a41bc3\": RecentStats: unable to find data in memory cache]" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.436383 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.448127 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.505190 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/extract/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.631549 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.747109 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.750942 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.776912 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:43:23 crc kubenswrapper[4970]: I1124 14:43:23.942601 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.001356 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.164366 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-utilities/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.386493 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-content/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.447672 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-utilities/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.475043 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-content/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.588223 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/registry-server/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.630246 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-content/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.648025 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/extract-utilities/0.log" Nov 24 14:43:24 crc kubenswrapper[4970]: I1124 14:43:24.835906 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.028739 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.043604 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.090808 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.138789 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bnws5_962a33e9-70d8-463f-b70b-c0427b287b5a/registry-server/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.230434 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.255233 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.267498 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/extract/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.429731 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8njc2_9776cf3e-bec6-4eca-874b-37e8dc5a755e/marketplace-operator/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.451286 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.625259 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.642094 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.653102 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.813186 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.852126 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:43:25 crc kubenswrapper[4970]: I1124 14:43:25.998491 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.088001 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/registry-server/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.174829 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.185076 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.230478 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.417235 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.430929 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:43:26 crc kubenswrapper[4970]: I1124 14:43:26.618983 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/registry-server/0.log" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.222764 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr"] Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223829 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223846 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223874 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223882 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223896 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223905 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223920 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223927 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223943 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223950 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="extract-utilities" Nov 24 14:45:00 crc kubenswrapper[4970]: E1124 14:45:00.223967 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.223974 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="extract-content" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.224209 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ba20f60-4645-4688-8665-d238986e0b8d" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.224225 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e2ce47-7b41-4e99-b2de-90b261ce0c21" containerName="registry-server" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.224985 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.233639 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.234347 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.245773 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr"] Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.323555 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzbmb\" (UniqueName: \"kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.323716 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.324311 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.426356 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.426464 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzbmb\" (UniqueName: \"kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.426510 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.427597 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.436356 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.456825 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzbmb\" (UniqueName: \"kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb\") pod \"collect-profiles-29399925-fzjdr\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.564244 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:00 crc kubenswrapper[4970]: I1124 14:45:00.916375 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr"] Nov 24 14:45:00 crc kubenswrapper[4970]: W1124 14:45:00.938810 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47e23df4_b182_4670_87a6_451c3c87d382.slice/crio-9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269 WatchSource:0}: Error finding container 9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269: Status 404 returned error can't find the container with id 9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269 Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.383933 4970 generic.go:334] "Generic (PLEG): container finished" podID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerID="cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f" exitCode=0 Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.384049 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" event={"ID":"e2aa957f-38ff-46f5-9e9b-a925278d9060","Type":"ContainerDied","Data":"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f"} Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.385010 4970 scope.go:117] "RemoveContainer" containerID="cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f" Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.385867 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" event={"ID":"47e23df4-b182-4670-87a6-451c3c87d382","Type":"ContainerStarted","Data":"c58793a6934d10d63977a14a9e87aa3dd9798086ede3df9f62873aa39f167fc8"} Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.385898 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" event={"ID":"47e23df4-b182-4670-87a6-451c3c87d382","Type":"ContainerStarted","Data":"9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269"} Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.443408 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" podStartSLOduration=1.443388377 podStartE2EDuration="1.443388377s" podCreationTimestamp="2025-11-24 14:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:45:01.430011719 +0000 UTC m=+3516.717769062" watchObservedRunningTime="2025-11-24 14:45:01.443388377 +0000 UTC m=+3516.731145680" Nov 24 14:45:01 crc kubenswrapper[4970]: I1124 14:45:01.495117 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqf9z_must-gather-vkhbq_e2aa957f-38ff-46f5-9e9b-a925278d9060/gather/0.log" Nov 24 14:45:02 crc kubenswrapper[4970]: I1124 14:45:02.397217 4970 generic.go:334] "Generic (PLEG): container finished" podID="47e23df4-b182-4670-87a6-451c3c87d382" containerID="c58793a6934d10d63977a14a9e87aa3dd9798086ede3df9f62873aa39f167fc8" exitCode=0 Nov 24 14:45:02 crc kubenswrapper[4970]: I1124 14:45:02.397285 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" event={"ID":"47e23df4-b182-4670-87a6-451c3c87d382","Type":"ContainerDied","Data":"c58793a6934d10d63977a14a9e87aa3dd9798086ede3df9f62873aa39f167fc8"} Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.752725 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.903567 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzbmb\" (UniqueName: \"kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb\") pod \"47e23df4-b182-4670-87a6-451c3c87d382\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.903742 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume\") pod \"47e23df4-b182-4670-87a6-451c3c87d382\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.903901 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume\") pod \"47e23df4-b182-4670-87a6-451c3c87d382\" (UID: \"47e23df4-b182-4670-87a6-451c3c87d382\") " Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.904880 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume" (OuterVolumeSpecName: "config-volume") pod "47e23df4-b182-4670-87a6-451c3c87d382" (UID: "47e23df4-b182-4670-87a6-451c3c87d382"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.911559 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb" (OuterVolumeSpecName: "kube-api-access-zzbmb") pod "47e23df4-b182-4670-87a6-451c3c87d382" (UID: "47e23df4-b182-4670-87a6-451c3c87d382"). InnerVolumeSpecName "kube-api-access-zzbmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:03 crc kubenswrapper[4970]: I1124 14:45:03.913757 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "47e23df4-b182-4670-87a6-451c3c87d382" (UID: "47e23df4-b182-4670-87a6-451c3c87d382"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.007096 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzbmb\" (UniqueName: \"kubernetes.io/projected/47e23df4-b182-4670-87a6-451c3c87d382-kube-api-access-zzbmb\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.007137 4970 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47e23df4-b182-4670-87a6-451c3c87d382-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.007149 4970 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47e23df4-b182-4670-87a6-451c3c87d382-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.418907 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" event={"ID":"47e23df4-b182-4670-87a6-451c3c87d382","Type":"ContainerDied","Data":"9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269"} Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.418969 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b65231065a090c6617a935a49fe3cf5043dfc29d4f0d7aee4c79ed3068f0269" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.419008 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399925-fzjdr" Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.838784 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7"] Nov 24 14:45:04 crc kubenswrapper[4970]: I1124 14:45:04.850372 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399880-zmbj7"] Nov 24 14:45:05 crc kubenswrapper[4970]: I1124 14:45:05.490048 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc36780e-1f96-45c0-a50b-fd9befc47823" path="/var/lib/kubelet/pods/cc36780e-1f96-45c0-a50b-fd9befc47823/volumes" Nov 24 14:45:09 crc kubenswrapper[4970]: I1124 14:45:09.625605 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wqf9z/must-gather-vkhbq"] Nov 24 14:45:09 crc kubenswrapper[4970]: I1124 14:45:09.626305 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="copy" containerID="cri-o://d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967" gracePeriod=2 Nov 24 14:45:09 crc kubenswrapper[4970]: I1124 14:45:09.637272 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wqf9z/must-gather-vkhbq"] Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.107346 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqf9z_must-gather-vkhbq_e2aa957f-38ff-46f5-9e9b-a925278d9060/copy/0.log" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.107911 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.242188 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hgvs\" (UniqueName: \"kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs\") pod \"e2aa957f-38ff-46f5-9e9b-a925278d9060\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.242411 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output\") pod \"e2aa957f-38ff-46f5-9e9b-a925278d9060\" (UID: \"e2aa957f-38ff-46f5-9e9b-a925278d9060\") " Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.252738 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs" (OuterVolumeSpecName: "kube-api-access-4hgvs") pod "e2aa957f-38ff-46f5-9e9b-a925278d9060" (UID: "e2aa957f-38ff-46f5-9e9b-a925278d9060"). InnerVolumeSpecName "kube-api-access-4hgvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.346081 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hgvs\" (UniqueName: \"kubernetes.io/projected/e2aa957f-38ff-46f5-9e9b-a925278d9060-kube-api-access-4hgvs\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.371653 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e2aa957f-38ff-46f5-9e9b-a925278d9060" (UID: "e2aa957f-38ff-46f5-9e9b-a925278d9060"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.448180 4970 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e2aa957f-38ff-46f5-9e9b-a925278d9060-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.500148 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wqf9z_must-gather-vkhbq_e2aa957f-38ff-46f5-9e9b-a925278d9060/copy/0.log" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.500642 4970 generic.go:334] "Generic (PLEG): container finished" podID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerID="d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967" exitCode=143 Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.500713 4970 scope.go:117] "RemoveContainer" containerID="d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.500746 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wqf9z/must-gather-vkhbq" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.530215 4970 scope.go:117] "RemoveContainer" containerID="cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.589744 4970 scope.go:117] "RemoveContainer" containerID="d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967" Nov 24 14:45:10 crc kubenswrapper[4970]: E1124 14:45:10.590672 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967\": container with ID starting with d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967 not found: ID does not exist" containerID="d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.590719 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967"} err="failed to get container status \"d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967\": rpc error: code = NotFound desc = could not find container \"d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967\": container with ID starting with d68509e62060b46d52f04faa5688cb354d4a824d481d852d88cef1ca2b35f967 not found: ID does not exist" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.590742 4970 scope.go:117] "RemoveContainer" containerID="cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f" Nov 24 14:45:10 crc kubenswrapper[4970]: E1124 14:45:10.591264 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f\": container with ID starting with cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f not found: ID does not exist" containerID="cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f" Nov 24 14:45:10 crc kubenswrapper[4970]: I1124 14:45:10.591564 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f"} err="failed to get container status \"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f\": rpc error: code = NotFound desc = could not find container \"cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f\": container with ID starting with cd297b6db2003b46acdcd6e126528c0a4a245aea23c00d5046c0ba1f6ba4e66f not found: ID does not exist" Nov 24 14:45:11 crc kubenswrapper[4970]: I1124 14:45:11.204995 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:45:11 crc kubenswrapper[4970]: I1124 14:45:11.205522 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:45:11 crc kubenswrapper[4970]: I1124 14:45:11.495354 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" path="/var/lib/kubelet/pods/e2aa957f-38ff-46f5-9e9b-a925278d9060/volumes" Nov 24 14:45:41 crc kubenswrapper[4970]: I1124 14:45:41.204017 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:45:41 crc kubenswrapper[4970]: I1124 14:45:41.204870 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:45:47 crc kubenswrapper[4970]: I1124 14:45:47.792520 4970 scope.go:117] "RemoveContainer" containerID="c83804b15721950704e8f7ab50433d08eedd48449afd5755fc56244e84fccf48" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.206468 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.207138 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.207214 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.208335 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.208433 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" gracePeriod=600 Nov 24 14:46:11 crc kubenswrapper[4970]: E1124 14:46:11.349231 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.594242 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" exitCode=0 Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.594281 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143"} Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.594315 4970 scope.go:117] "RemoveContainer" containerID="736429c67b06c922b52a18097c939e826265dcb13a192adbf03557d7a5ad26a8" Nov 24 14:46:11 crc kubenswrapper[4970]: I1124 14:46:11.594680 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:46:11 crc kubenswrapper[4970]: E1124 14:46:11.595605 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:46:25 crc kubenswrapper[4970]: I1124 14:46:25.483303 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:46:25 crc kubenswrapper[4970]: E1124 14:46:25.484784 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:46:38 crc kubenswrapper[4970]: I1124 14:46:38.470996 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:46:38 crc kubenswrapper[4970]: E1124 14:46:38.472087 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:46:50 crc kubenswrapper[4970]: I1124 14:46:50.471542 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:46:50 crc kubenswrapper[4970]: E1124 14:46:50.472889 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:03 crc kubenswrapper[4970]: I1124 14:47:03.471825 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:47:03 crc kubenswrapper[4970]: E1124 14:47:03.473095 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:17 crc kubenswrapper[4970]: I1124 14:47:17.471367 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:47:17 crc kubenswrapper[4970]: E1124 14:47:17.472506 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:28 crc kubenswrapper[4970]: I1124 14:47:28.470988 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:47:28 crc kubenswrapper[4970]: E1124 14:47:28.472107 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:40 crc kubenswrapper[4970]: I1124 14:47:40.472460 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:47:40 crc kubenswrapper[4970]: E1124 14:47:40.473911 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.977510 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gvrb/must-gather-9mwph"] Nov 24 14:47:46 crc kubenswrapper[4970]: E1124 14:47:46.978310 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e23df4-b182-4670-87a6-451c3c87d382" containerName="collect-profiles" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978323 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e23df4-b182-4670-87a6-451c3c87d382" containerName="collect-profiles" Nov 24 14:47:46 crc kubenswrapper[4970]: E1124 14:47:46.978338 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="gather" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978344 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="gather" Nov 24 14:47:46 crc kubenswrapper[4970]: E1124 14:47:46.978376 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="copy" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978381 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="copy" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978551 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="copy" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978564 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e23df4-b182-4670-87a6-451c3c87d382" containerName="collect-profiles" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.978591 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2aa957f-38ff-46f5-9e9b-a925278d9060" containerName="gather" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.979504 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.981778 4970 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4gvrb"/"default-dockercfg-wdhh2" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.981796 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gvrb"/"openshift-service-ca.crt" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.981853 4970 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4gvrb"/"kube-root-ca.crt" Nov 24 14:47:46 crc kubenswrapper[4970]: I1124 14:47:46.986855 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gvrb/must-gather-9mwph"] Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.089412 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtpj2\" (UniqueName: \"kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.089514 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.191464 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtpj2\" (UniqueName: \"kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.191598 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.192016 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.214106 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtpj2\" (UniqueName: \"kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2\") pod \"must-gather-9mwph\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.300616 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.842715 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4gvrb/must-gather-9mwph"] Nov 24 14:47:47 crc kubenswrapper[4970]: I1124 14:47:47.870403 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/must-gather-9mwph" event={"ID":"f26a4374-de0f-42db-abdc-3cbec3f634a0","Type":"ContainerStarted","Data":"c6fb917c1baea0a7aa4541c20dc34b36b5cbea69160e443e152f490b89be2749"} Nov 24 14:47:48 crc kubenswrapper[4970]: I1124 14:47:48.882708 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/must-gather-9mwph" event={"ID":"f26a4374-de0f-42db-abdc-3cbec3f634a0","Type":"ContainerStarted","Data":"97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713"} Nov 24 14:47:48 crc kubenswrapper[4970]: I1124 14:47:48.883165 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/must-gather-9mwph" event={"ID":"f26a4374-de0f-42db-abdc-3cbec3f634a0","Type":"ContainerStarted","Data":"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859"} Nov 24 14:47:48 crc kubenswrapper[4970]: I1124 14:47:48.906534 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gvrb/must-gather-9mwph" podStartSLOduration=2.906506431 podStartE2EDuration="2.906506431s" podCreationTimestamp="2025-11-24 14:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:47:48.904023933 +0000 UTC m=+3684.191781216" watchObservedRunningTime="2025-11-24 14:47:48.906506431 +0000 UTC m=+3684.194263734" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.879927 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-bc4sn"] Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.881407 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.884874 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4fgq\" (UniqueName: \"kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.884916 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.986909 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4fgq\" (UniqueName: \"kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.987147 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:51 crc kubenswrapper[4970]: I1124 14:47:51.987354 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.004592 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4fgq\" (UniqueName: \"kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq\") pod \"crc-debug-bc4sn\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.201602 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.470154 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:47:52 crc kubenswrapper[4970]: E1124 14:47:52.470720 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.924823 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" event={"ID":"9662890c-373a-4671-b2cb-4b932f841deb","Type":"ContainerStarted","Data":"6d74fc5014135bce3330c859da4e325645a2a2ede4ec148a37cd4174b34a349b"} Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.925129 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" event={"ID":"9662890c-373a-4671-b2cb-4b932f841deb","Type":"ContainerStarted","Data":"b12eeedbf8c9835afcfdc3a8da62d5ec0abcdf66c2a7ae52f2a4935064f5a773"} Nov 24 14:47:52 crc kubenswrapper[4970]: I1124 14:47:52.942188 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" podStartSLOduration=1.942171916 podStartE2EDuration="1.942171916s" podCreationTimestamp="2025-11-24 14:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 14:47:52.936627644 +0000 UTC m=+3688.224384937" watchObservedRunningTime="2025-11-24 14:47:52.942171916 +0000 UTC m=+3688.229929209" Nov 24 14:48:05 crc kubenswrapper[4970]: I1124 14:48:05.485926 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:48:05 crc kubenswrapper[4970]: E1124 14:48:05.487062 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.184439 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.192968 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.202772 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.271933 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.272007 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcljg\" (UniqueName: \"kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.272029 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.373314 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.373359 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcljg\" (UniqueName: \"kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.373512 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.373885 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.373910 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.392744 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcljg\" (UniqueName: \"kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg\") pod \"redhat-operators-kw7wf\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.509631 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:09 crc kubenswrapper[4970]: I1124 14:48:09.998750 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:10 crc kubenswrapper[4970]: W1124 14:48:10.012827 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0082e1cb_32ed_4af6_a965_bec253af0fda.slice/crio-70a3c0cde54da0acb42f3fde71492839e8cc390c0e34803dac02e6f9efd0eb58 WatchSource:0}: Error finding container 70a3c0cde54da0acb42f3fde71492839e8cc390c0e34803dac02e6f9efd0eb58: Status 404 returned error can't find the container with id 70a3c0cde54da0acb42f3fde71492839e8cc390c0e34803dac02e6f9efd0eb58 Nov 24 14:48:10 crc kubenswrapper[4970]: I1124 14:48:10.071787 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerStarted","Data":"70a3c0cde54da0acb42f3fde71492839e8cc390c0e34803dac02e6f9efd0eb58"} Nov 24 14:48:11 crc kubenswrapper[4970]: I1124 14:48:11.081963 4970 generic.go:334] "Generic (PLEG): container finished" podID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerID="d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055" exitCode=0 Nov 24 14:48:11 crc kubenswrapper[4970]: I1124 14:48:11.082232 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerDied","Data":"d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055"} Nov 24 14:48:11 crc kubenswrapper[4970]: I1124 14:48:11.084940 4970 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 14:48:13 crc kubenswrapper[4970]: I1124 14:48:13.101968 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerStarted","Data":"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea"} Nov 24 14:48:15 crc kubenswrapper[4970]: I1124 14:48:15.119489 4970 generic.go:334] "Generic (PLEG): container finished" podID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerID="701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea" exitCode=0 Nov 24 14:48:15 crc kubenswrapper[4970]: I1124 14:48:15.120084 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerDied","Data":"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea"} Nov 24 14:48:16 crc kubenswrapper[4970]: I1124 14:48:16.131132 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerStarted","Data":"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f"} Nov 24 14:48:19 crc kubenswrapper[4970]: I1124 14:48:19.470126 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:48:19 crc kubenswrapper[4970]: E1124 14:48:19.470632 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:48:19 crc kubenswrapper[4970]: I1124 14:48:19.509942 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:19 crc kubenswrapper[4970]: I1124 14:48:19.510000 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:20 crc kubenswrapper[4970]: I1124 14:48:20.582842 4970 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kw7wf" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="registry-server" probeResult="failure" output=< Nov 24 14:48:20 crc kubenswrapper[4970]: timeout: failed to connect service ":50051" within 1s Nov 24 14:48:20 crc kubenswrapper[4970]: > Nov 24 14:48:24 crc kubenswrapper[4970]: I1124 14:48:24.202490 4970 generic.go:334] "Generic (PLEG): container finished" podID="9662890c-373a-4671-b2cb-4b932f841deb" containerID="6d74fc5014135bce3330c859da4e325645a2a2ede4ec148a37cd4174b34a349b" exitCode=0 Nov 24 14:48:24 crc kubenswrapper[4970]: I1124 14:48:24.202596 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" event={"ID":"9662890c-373a-4671-b2cb-4b932f841deb","Type":"ContainerDied","Data":"6d74fc5014135bce3330c859da4e325645a2a2ede4ec148a37cd4174b34a349b"} Nov 24 14:48:24 crc kubenswrapper[4970]: I1124 14:48:24.218777 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kw7wf" podStartSLOduration=10.741571782 podStartE2EDuration="15.218760263s" podCreationTimestamp="2025-11-24 14:48:09 +0000 UTC" firstStartedPulling="2025-11-24 14:48:11.084689699 +0000 UTC m=+3706.372447002" lastFinishedPulling="2025-11-24 14:48:15.56187816 +0000 UTC m=+3710.849635483" observedRunningTime="2025-11-24 14:48:16.164606785 +0000 UTC m=+3711.452364078" watchObservedRunningTime="2025-11-24 14:48:24.218760263 +0000 UTC m=+3719.506517556" Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.308375 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.349803 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-bc4sn"] Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.358978 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-bc4sn"] Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.489896 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host\") pod \"9662890c-373a-4671-b2cb-4b932f841deb\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.490004 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host" (OuterVolumeSpecName: "host") pod "9662890c-373a-4671-b2cb-4b932f841deb" (UID: "9662890c-373a-4671-b2cb-4b932f841deb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.490049 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4fgq\" (UniqueName: \"kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq\") pod \"9662890c-373a-4671-b2cb-4b932f841deb\" (UID: \"9662890c-373a-4671-b2cb-4b932f841deb\") " Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.490451 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9662890c-373a-4671-b2cb-4b932f841deb-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.497510 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq" (OuterVolumeSpecName: "kube-api-access-g4fgq") pod "9662890c-373a-4671-b2cb-4b932f841deb" (UID: "9662890c-373a-4671-b2cb-4b932f841deb"). InnerVolumeSpecName "kube-api-access-g4fgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:48:25 crc kubenswrapper[4970]: I1124 14:48:25.592623 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4fgq\" (UniqueName: \"kubernetes.io/projected/9662890c-373a-4671-b2cb-4b932f841deb-kube-api-access-g4fgq\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.231945 4970 scope.go:117] "RemoveContainer" containerID="6d74fc5014135bce3330c859da4e325645a2a2ede4ec148a37cd4174b34a349b" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.233147 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-bc4sn" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.537095 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-52sst"] Nov 24 14:48:26 crc kubenswrapper[4970]: E1124 14:48:26.537563 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9662890c-373a-4671-b2cb-4b932f841deb" containerName="container-00" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.537596 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="9662890c-373a-4671-b2cb-4b932f841deb" containerName="container-00" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.537812 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="9662890c-373a-4671-b2cb-4b932f841deb" containerName="container-00" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.538410 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.710874 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6ktw\" (UniqueName: \"kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.711225 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.813065 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6ktw\" (UniqueName: \"kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.813193 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.813329 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.835598 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6ktw\" (UniqueName: \"kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw\") pod \"crc-debug-52sst\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: I1124 14:48:26.856918 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:26 crc kubenswrapper[4970]: W1124 14:48:26.885111 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70c79570_7e71_4a00_a1d0_288124050a93.slice/crio-93d990490cde10c42304679805dc6b8ec77047b90048247aa9308847d7d3520f WatchSource:0}: Error finding container 93d990490cde10c42304679805dc6b8ec77047b90048247aa9308847d7d3520f: Status 404 returned error can't find the container with id 93d990490cde10c42304679805dc6b8ec77047b90048247aa9308847d7d3520f Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.244115 4970 generic.go:334] "Generic (PLEG): container finished" podID="70c79570-7e71-4a00-a1d0-288124050a93" containerID="93b8397f6128352afb8f3fb239c1f6362845cef937ca55a6502ce898d308a0dc" exitCode=0 Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.244229 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-52sst" event={"ID":"70c79570-7e71-4a00-a1d0-288124050a93","Type":"ContainerDied","Data":"93b8397f6128352afb8f3fb239c1f6362845cef937ca55a6502ce898d308a0dc"} Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.244636 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-52sst" event={"ID":"70c79570-7e71-4a00-a1d0-288124050a93","Type":"ContainerStarted","Data":"93d990490cde10c42304679805dc6b8ec77047b90048247aa9308847d7d3520f"} Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.482952 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9662890c-373a-4671-b2cb-4b932f841deb" path="/var/lib/kubelet/pods/9662890c-373a-4671-b2cb-4b932f841deb/volumes" Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.928757 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-52sst"] Nov 24 14:48:27 crc kubenswrapper[4970]: I1124 14:48:27.942251 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-52sst"] Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.385206 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.527012 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6ktw\" (UniqueName: \"kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw\") pod \"70c79570-7e71-4a00-a1d0-288124050a93\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.527208 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host\") pod \"70c79570-7e71-4a00-a1d0-288124050a93\" (UID: \"70c79570-7e71-4a00-a1d0-288124050a93\") " Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.527278 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host" (OuterVolumeSpecName: "host") pod "70c79570-7e71-4a00-a1d0-288124050a93" (UID: "70c79570-7e71-4a00-a1d0-288124050a93"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.528066 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70c79570-7e71-4a00-a1d0-288124050a93-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.534043 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw" (OuterVolumeSpecName: "kube-api-access-l6ktw") pod "70c79570-7e71-4a00-a1d0-288124050a93" (UID: "70c79570-7e71-4a00-a1d0-288124050a93"). InnerVolumeSpecName "kube-api-access-l6ktw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:48:28 crc kubenswrapper[4970]: I1124 14:48:28.629325 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6ktw\" (UniqueName: \"kubernetes.io/projected/70c79570-7e71-4a00-a1d0-288124050a93-kube-api-access-l6ktw\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.171962 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-cdlxs"] Nov 24 14:48:29 crc kubenswrapper[4970]: E1124 14:48:29.172497 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c79570-7e71-4a00-a1d0-288124050a93" containerName="container-00" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.172516 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c79570-7e71-4a00-a1d0-288124050a93" containerName="container-00" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.172865 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c79570-7e71-4a00-a1d0-288124050a93" containerName="container-00" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.173828 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.242551 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.242692 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw26m\" (UniqueName: \"kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.302885 4970 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93d990490cde10c42304679805dc6b8ec77047b90048247aa9308847d7d3520f" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.302961 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-52sst" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.344899 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.344969 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw26m\" (UniqueName: \"kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.345550 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.371991 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw26m\" (UniqueName: \"kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m\") pod \"crc-debug-cdlxs\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.485043 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c79570-7e71-4a00-a1d0-288124050a93" path="/var/lib/kubelet/pods/70c79570-7e71-4a00-a1d0-288124050a93/volumes" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.496393 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:29 crc kubenswrapper[4970]: W1124 14:48:29.561448 4970 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2767a8b2_89be_48b9_976b_839a367f78fd.slice/crio-51b90b6153afb9d12d08bc747ffd035f3db87202d09f03ba00896640823b7cd1 WatchSource:0}: Error finding container 51b90b6153afb9d12d08bc747ffd035f3db87202d09f03ba00896640823b7cd1: Status 404 returned error can't find the container with id 51b90b6153afb9d12d08bc747ffd035f3db87202d09f03ba00896640823b7cd1 Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.621478 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.691109 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:29 crc kubenswrapper[4970]: I1124 14:48:29.881460 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.318785 4970 generic.go:334] "Generic (PLEG): container finished" podID="2767a8b2-89be-48b9-976b-839a367f78fd" containerID="c9a8d4dcc1ebbe6c0ab33ed31a5fc7f5eff48a2ae8b784f9de90f85071f582c5" exitCode=0 Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.318861 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" event={"ID":"2767a8b2-89be-48b9-976b-839a367f78fd","Type":"ContainerDied","Data":"c9a8d4dcc1ebbe6c0ab33ed31a5fc7f5eff48a2ae8b784f9de90f85071f582c5"} Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.318938 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" event={"ID":"2767a8b2-89be-48b9-976b-839a367f78fd","Type":"ContainerStarted","Data":"51b90b6153afb9d12d08bc747ffd035f3db87202d09f03ba00896640823b7cd1"} Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.373419 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-cdlxs"] Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.379005 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gvrb/crc-debug-cdlxs"] Nov 24 14:48:30 crc kubenswrapper[4970]: I1124 14:48:30.470469 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:48:30 crc kubenswrapper[4970]: E1124 14:48:30.470818 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.338657 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kw7wf" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="registry-server" containerID="cri-o://1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f" gracePeriod=2 Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.620200 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.714651 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw26m\" (UniqueName: \"kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m\") pod \"2767a8b2-89be-48b9-976b-839a367f78fd\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.715800 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host\") pod \"2767a8b2-89be-48b9-976b-839a367f78fd\" (UID: \"2767a8b2-89be-48b9-976b-839a367f78fd\") " Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.715972 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host" (OuterVolumeSpecName: "host") pod "2767a8b2-89be-48b9-976b-839a367f78fd" (UID: "2767a8b2-89be-48b9-976b-839a367f78fd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.716797 4970 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2767a8b2-89be-48b9-976b-839a367f78fd-host\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.729840 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m" (OuterVolumeSpecName: "kube-api-access-rw26m") pod "2767a8b2-89be-48b9-976b-839a367f78fd" (UID: "2767a8b2-89be-48b9-976b-839a367f78fd"). InnerVolumeSpecName "kube-api-access-rw26m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.817530 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw26m\" (UniqueName: \"kubernetes.io/projected/2767a8b2-89be-48b9-976b-839a367f78fd-kube-api-access-rw26m\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.848737 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.918401 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcljg\" (UniqueName: \"kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg\") pod \"0082e1cb-32ed-4af6-a965-bec253af0fda\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.918852 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content\") pod \"0082e1cb-32ed-4af6-a965-bec253af0fda\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.919127 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities\") pod \"0082e1cb-32ed-4af6-a965-bec253af0fda\" (UID: \"0082e1cb-32ed-4af6-a965-bec253af0fda\") " Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.919932 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities" (OuterVolumeSpecName: "utilities") pod "0082e1cb-32ed-4af6-a965-bec253af0fda" (UID: "0082e1cb-32ed-4af6-a965-bec253af0fda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:48:31 crc kubenswrapper[4970]: I1124 14:48:31.924538 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg" (OuterVolumeSpecName: "kube-api-access-hcljg") pod "0082e1cb-32ed-4af6-a965-bec253af0fda" (UID: "0082e1cb-32ed-4af6-a965-bec253af0fda"). InnerVolumeSpecName "kube-api-access-hcljg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.022027 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.022084 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcljg\" (UniqueName: \"kubernetes.io/projected/0082e1cb-32ed-4af6-a965-bec253af0fda-kube-api-access-hcljg\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.034505 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0082e1cb-32ed-4af6-a965-bec253af0fda" (UID: "0082e1cb-32ed-4af6-a965-bec253af0fda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.123334 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0082e1cb-32ed-4af6-a965-bec253af0fda-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.354894 4970 scope.go:117] "RemoveContainer" containerID="c9a8d4dcc1ebbe6c0ab33ed31a5fc7f5eff48a2ae8b784f9de90f85071f582c5" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.354943 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/crc-debug-cdlxs" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.359812 4970 generic.go:334] "Generic (PLEG): container finished" podID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerID="1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f" exitCode=0 Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.359884 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerDied","Data":"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f"} Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.359931 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kw7wf" event={"ID":"0082e1cb-32ed-4af6-a965-bec253af0fda","Type":"ContainerDied","Data":"70a3c0cde54da0acb42f3fde71492839e8cc390c0e34803dac02e6f9efd0eb58"} Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.360033 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kw7wf" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.471508 4970 scope.go:117] "RemoveContainer" containerID="1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.502414 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.512717 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kw7wf"] Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.514288 4970 scope.go:117] "RemoveContainer" containerID="701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.545881 4970 scope.go:117] "RemoveContainer" containerID="d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.586018 4970 scope.go:117] "RemoveContainer" containerID="1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f" Nov 24 14:48:32 crc kubenswrapper[4970]: E1124 14:48:32.586476 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f\": container with ID starting with 1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f not found: ID does not exist" containerID="1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.586528 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f"} err="failed to get container status \"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f\": rpc error: code = NotFound desc = could not find container \"1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f\": container with ID starting with 1fe635e8afbc2f3f86e92e16f6e6d9c895585e0f3de10fb86a8c761ba46d067f not found: ID does not exist" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.586559 4970 scope.go:117] "RemoveContainer" containerID="701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea" Nov 24 14:48:32 crc kubenswrapper[4970]: E1124 14:48:32.587167 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea\": container with ID starting with 701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea not found: ID does not exist" containerID="701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.587200 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea"} err="failed to get container status \"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea\": rpc error: code = NotFound desc = could not find container \"701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea\": container with ID starting with 701651bd502285f6913b7850e2e95795a4b070a9e28b8fd0b667f9ff546b56ea not found: ID does not exist" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.587224 4970 scope.go:117] "RemoveContainer" containerID="d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055" Nov 24 14:48:32 crc kubenswrapper[4970]: E1124 14:48:32.587510 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055\": container with ID starting with d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055 not found: ID does not exist" containerID="d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055" Nov 24 14:48:32 crc kubenswrapper[4970]: I1124 14:48:32.587542 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055"} err="failed to get container status \"d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055\": rpc error: code = NotFound desc = could not find container \"d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055\": container with ID starting with d6e6af5e35006189bac6e35d0cdf2da138f73c560559f559a538dbdea7370055 not found: ID does not exist" Nov 24 14:48:33 crc kubenswrapper[4970]: I1124 14:48:33.500635 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" path="/var/lib/kubelet/pods/0082e1cb-32ed-4af6-a965-bec253af0fda/volumes" Nov 24 14:48:33 crc kubenswrapper[4970]: I1124 14:48:33.501422 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2767a8b2-89be-48b9-976b-839a367f78fd" path="/var/lib/kubelet/pods/2767a8b2-89be-48b9-976b-839a367f78fd/volumes" Nov 24 14:48:42 crc kubenswrapper[4970]: I1124 14:48:42.470156 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:48:42 crc kubenswrapper[4970]: E1124 14:48:42.470989 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:48:56 crc kubenswrapper[4970]: I1124 14:48:56.469849 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:48:56 crc kubenswrapper[4970]: E1124 14:48:56.470553 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:48:56 crc kubenswrapper[4970]: I1124 14:48:56.862330 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756b7f5776-9gksb_83e2deca-2802-4021-9e47-b933ace28f4a/barbican-api/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.072666 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756b7f5776-9gksb_83e2deca-2802-4021-9e47-b933ace28f4a/barbican-api-log/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.122741 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-688cdc6cb6-fkrjb_df1d70bf-d612-4ef1-a113-6be419867e1d/barbican-keystone-listener/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.147041 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-688cdc6cb6-fkrjb_df1d70bf-d612-4ef1-a113-6be419867e1d/barbican-keystone-listener-log/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.306355 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-85596cc87-lvc2b_17b7d151-f71d-4f02-9d52-4c7ba81ab89e/barbican-worker-log/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.320861 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-85596cc87-lvc2b_17b7d151-f71d-4f02-9d52-4c7ba81ab89e/barbican-worker/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.495189 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k8tcv_b9700095-d8fb-496e-981d-de1e30ad7e8a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.558500 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/ceilometer-central-agent/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.628343 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/ceilometer-notification-agent/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.704298 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/sg-core/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.708336 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_de278d27-bc30-4d46-a4b3-3019fb4e6e3c/proxy-httpd/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.837484 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9c6c4a13-7335-4e6f-b424-90a6a003a894/cinder-api/0.log" Nov 24 14:48:57 crc kubenswrapper[4970]: I1124 14:48:57.877230 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9c6c4a13-7335-4e6f-b424-90a6a003a894/cinder-api-log/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.078905 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c8304c7a-67d5-4d0f-afce-231f09bf48ea/cinder-scheduler/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.124562 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c8304c7a-67d5-4d0f-afce-231f09bf48ea/probe/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.245988 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-s49zl_34ffc82b-5c7e-4995-8dab-f09ed6e1c2ad/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.312201 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gdg4z_35b1536c-4ac4-42f5-9f1f-09838c75d1b7/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.447844 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/init/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.605964 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/init/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.667697 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d5cf5b645-zwshh_83916562-6f47-4b0b-8d48-6e48f0104ae6/dnsmasq-dns/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.724219 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-7lhq8_bdec1f38-8d93-4298-ad14-872b58cff4e9/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.840319 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0713092e-39d2-40e1-b708-a2e2addd8249/glance-httpd/0.log" Nov 24 14:48:58 crc kubenswrapper[4970]: I1124 14:48:58.867979 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0713092e-39d2-40e1-b708-a2e2addd8249/glance-log/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.027333 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_639b382e-fe48-4188-ba6b-1b2c5688898c/glance-httpd/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.073309 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_639b382e-fe48-4188-ba6b-1b2c5688898c/glance-log/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.197203 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bb8f9fccd-2m5xk_070dfeee-2ee7-4b9e-bdbf-7320a48a0659/horizon/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.387992 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kpx5p_a87a7cba-b0be-443b-8393-6f0f21e1ccb9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.509981 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fctzs_50a618ec-6a8b-4234-8000-253f692b204a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.566250 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bb8f9fccd-2m5xk_070dfeee-2ee7-4b9e-bdbf-7320a48a0659/horizon-log/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.744477 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3753fd35-b151-4bde-aa8c-a9c5f0a65f3f/kube-state-metrics/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.775278 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8c975fcc4-8cmzb_24ee1a5e-62c4-42f5-999b-27ddf44b33ae/keystone-api/0.log" Nov 24 14:48:59 crc kubenswrapper[4970]: I1124 14:48:59.992324 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-sk2bv_5dcb0227-fbb2-492f-819b-02dd5de1a728/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:00 crc kubenswrapper[4970]: I1124 14:49:00.339950 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6464bbf59c-l44tb_ca5d8b57-826f-4811-a8b4-4f48e3082473/neutron-httpd/0.log" Nov 24 14:49:00 crc kubenswrapper[4970]: I1124 14:49:00.383812 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6464bbf59c-l44tb_ca5d8b57-826f-4811-a8b4-4f48e3082473/neutron-api/0.log" Nov 24 14:49:00 crc kubenswrapper[4970]: I1124 14:49:00.571011 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-jgvnw_c1a37385-6027-4ee8-ad98-269f93d2c46d/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.041360 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_05e7b6fd-92df-4c42-b84e-151948459084/nova-api-log/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.110012 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e8645891-3d8f-4506-8139-d13f7ee73054/nova-cell0-conductor-conductor/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.364146 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_05e7b6fd-92df-4c42-b84e-151948459084/nova-api-api/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.367188 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0c35af6c-0331-4c98-9cb4-116c2ed45ec7/nova-cell1-conductor-conductor/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.463149 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5271e19c-5bd9-4c25-a6b9-a919f1445cca/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.696239 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-xmx7z_cfcd6f60-0b2f-4066-bb72-9634b21eb168/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:01 crc kubenswrapper[4970]: I1124 14:49:01.781749 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9244354d-e3fc-411a-bdc4-2d81cc2b72a0/nova-metadata-log/0.log" Nov 24 14:49:02 crc kubenswrapper[4970]: I1124 14:49:02.108785 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/mysql-bootstrap/0.log" Nov 24 14:49:02 crc kubenswrapper[4970]: I1124 14:49:02.113312 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_92b5c70e-7c5c-465f-a44c-adc0aed6c664/nova-scheduler-scheduler/0.log" Nov 24 14:49:02 crc kubenswrapper[4970]: I1124 14:49:02.233860 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/mysql-bootstrap/0.log" Nov 24 14:49:02 crc kubenswrapper[4970]: I1124 14:49:02.286479 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_69fa6c84-911e-41f0-904d-1bd1b08d3b0c/galera/0.log" Nov 24 14:49:02 crc kubenswrapper[4970]: I1124 14:49:02.421798 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/mysql-bootstrap/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.000591 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9244354d-e3fc-411a-bdc4-2d81cc2b72a0/nova-metadata-metadata/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.037495 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ghss2"] Nov 24 14:49:03 crc kubenswrapper[4970]: E1124 14:49:03.038431 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="registry-server" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038454 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="registry-server" Nov 24 14:49:03 crc kubenswrapper[4970]: E1124 14:49:03.038465 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2767a8b2-89be-48b9-976b-839a367f78fd" containerName="container-00" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038472 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="2767a8b2-89be-48b9-976b-839a367f78fd" containerName="container-00" Nov 24 14:49:03 crc kubenswrapper[4970]: E1124 14:49:03.038486 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="extract-content" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038492 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="extract-content" Nov 24 14:49:03 crc kubenswrapper[4970]: E1124 14:49:03.038511 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="extract-utilities" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038517 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="extract-utilities" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038760 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="2767a8b2-89be-48b9-976b-839a367f78fd" containerName="container-00" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.038784 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="0082e1cb-32ed-4af6-a965-bec253af0fda" containerName="registry-server" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.040108 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.064491 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ghss2"] Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.064706 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/mysql-bootstrap/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.148814 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-kube-api-access-ffwm2\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.148854 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-catalog-content\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.148877 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-utilities\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.168748 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f0b6f151-1b32-4346-aabb-38ce2d89f50f/galera/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.241627 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0c4f17a4-9a8d-4824-8370-6aea04d166ae/openstackclient/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.251378 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-kube-api-access-ffwm2\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.251420 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-catalog-content\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.251447 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-utilities\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.251964 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-utilities\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.252424 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-catalog-content\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.270230 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffwm2\" (UniqueName: \"kubernetes.io/projected/8d0b0fbd-8b3b-4faa-90f8-02a413146f8a-kube-api-access-ffwm2\") pod \"community-operators-ghss2\" (UID: \"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a\") " pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.363512 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.587378 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-78b67_bb1eaff5-0428-4ab0-ae3b-21f192c95279/ovn-controller/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.627222 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-vz6qv_3f1018f7-87e9-49d7-aa02-ee711b839d82/openstack-network-exporter/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.865659 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server-init/0.log" Nov 24 14:49:03 crc kubenswrapper[4970]: I1124 14:49:03.939464 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ghss2"] Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.039202 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.071709 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovs-vswitchd/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.088361 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r89nk_c5806c43-906c-4790-82e3-511bd817eb88/ovsdb-server-init/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.664167 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8561d773-9da8-4031-92c2-84ebd42a1afc/ovn-northd/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.689324 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_8561d773-9da8-4031-92c2-84ebd42a1afc/openstack-network-exporter/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.698019 4970 generic.go:334] "Generic (PLEG): container finished" podID="8d0b0fbd-8b3b-4faa-90f8-02a413146f8a" containerID="00e8041529c5178d298f40fd2fdec97b1cc26206da9bba98dab191cd2d20201d" exitCode=0 Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.698064 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ghss2" event={"ID":"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a","Type":"ContainerDied","Data":"00e8041529c5178d298f40fd2fdec97b1cc26206da9bba98dab191cd2d20201d"} Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.698091 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ghss2" event={"ID":"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a","Type":"ContainerStarted","Data":"1d38dc780bcbc55b47c25a5283015493b8db26d24d9a8a236bd028ed3f6bc783"} Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.711480 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7dr7t_1843fb9f-02d8-4021-8ef5-f5416f439ef5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.901988 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90130f89-b1ad-43b9-9ab4-296530171bd6/openstack-network-exporter/0.log" Nov 24 14:49:04 crc kubenswrapper[4970]: I1124 14:49:04.926312 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90130f89-b1ad-43b9-9ab4-296530171bd6/ovsdbserver-nb/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.053474 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4910fec1-46d5-419d-baf9-3fa44a49b1d5/openstack-network-exporter/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.112484 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4910fec1-46d5-419d-baf9-3fa44a49b1d5/ovsdbserver-sb/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.358713 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd4546cfb-h57bs_12d2adeb-5c77-483a-bb89-63b7a50c4935/placement-api/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.401608 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd4546cfb-h57bs_12d2adeb-5c77-483a-bb89-63b7a50c4935/placement-log/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.437046 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/setup-container/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.619245 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/setup-container/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.664175 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9f4db0a4-ca88-4271-84f8-8c917057e031/rabbitmq/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.685677 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/setup-container/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.912136 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-qqv9m_ee264c1e-7f6a-4dc1-b406-aced8f917394/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.916497 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/setup-container/0.log" Nov 24 14:49:05 crc kubenswrapper[4970]: I1124 14:49:05.941018 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_6e9054df-765d-49f9-a41e-5bc3fb732691/rabbitmq/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.078133 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-c9km7_f6e35ba6-d32d-4204-a731-7f0e24802382/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.137531 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-glq2r_dafb717b-e6d5-41f4-8dab-acbd1db9f563/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.309242 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zb64x_03e03094-db8d-4000-84ea-23ace55b1768/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.449210 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-pxxvx_6ac1b7b7-0e52-40cc-bd5e-042430621073/ssh-known-hosts-edpm-deployment/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.623678 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76bd896ccf-szjxt_32811b6b-86a2-4762-94fd-2eef9eedad76/proxy-server/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.706329 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t72v9_b596f962-fb31-4389-937e-f42750aeaf6a/swift-ring-rebalance/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.718803 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-76bd896ccf-szjxt_32811b6b-86a2-4762-94fd-2eef9eedad76/proxy-httpd/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.887442 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-reaper/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.903606 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-auditor/0.log" Nov 24 14:49:06 crc kubenswrapper[4970]: I1124 14:49:06.940996 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-replicator/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.048622 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/account-server/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.063799 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-auditor/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.180129 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-server/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.182121 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-replicator/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.213193 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/container-updater/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.299074 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-auditor/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.383387 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-expirer/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.420597 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-server/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.423585 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-replicator/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.470110 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:49:07 crc kubenswrapper[4970]: E1124 14:49:07.470492 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.518614 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/object-updater/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.533030 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/rsync/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.643420 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_1fa4b14f-acb6-40eb-90ba-be3007901b3a/swift-recon-cron/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.755515 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-whzwn_196cf6ec-261f-4c0e-87f4-30c0e60e7bd2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:07 crc kubenswrapper[4970]: I1124 14:49:07.875829 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_1ffd9b46-906d-47a2-870e-6ad3d5381c2d/tempest-tests-tempest-tests-runner/0.log" Nov 24 14:49:08 crc kubenswrapper[4970]: I1124 14:49:08.186329 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_41506018-039f-4567-8ed7-2501e5d565a6/test-operator-logs-container/0.log" Nov 24 14:49:08 crc kubenswrapper[4970]: I1124 14:49:08.330076 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2nk8w_4665b179-69e1-41a9-a621-5681262a765a/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 14:49:11 crc kubenswrapper[4970]: I1124 14:49:11.780942 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ghss2" event={"ID":"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a","Type":"ContainerStarted","Data":"2ff7adce3cf33c238ae25dd1ea513253136c2cc84afdbfa8767bf621794d9707"} Nov 24 14:49:12 crc kubenswrapper[4970]: I1124 14:49:12.790984 4970 generic.go:334] "Generic (PLEG): container finished" podID="8d0b0fbd-8b3b-4faa-90f8-02a413146f8a" containerID="2ff7adce3cf33c238ae25dd1ea513253136c2cc84afdbfa8767bf621794d9707" exitCode=0 Nov 24 14:49:12 crc kubenswrapper[4970]: I1124 14:49:12.791050 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ghss2" event={"ID":"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a","Type":"ContainerDied","Data":"2ff7adce3cf33c238ae25dd1ea513253136c2cc84afdbfa8767bf621794d9707"} Nov 24 14:49:13 crc kubenswrapper[4970]: I1124 14:49:13.800738 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ghss2" event={"ID":"8d0b0fbd-8b3b-4faa-90f8-02a413146f8a","Type":"ContainerStarted","Data":"846d130ea67f6fdc483bb6625c24e33942a4ee6ba7330346492bb907b1ce69db"} Nov 24 14:49:13 crc kubenswrapper[4970]: I1124 14:49:13.825016 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ghss2" podStartSLOduration=2.216871905 podStartE2EDuration="10.825000061s" podCreationTimestamp="2025-11-24 14:49:03 +0000 UTC" firstStartedPulling="2025-11-24 14:49:04.699816587 +0000 UTC m=+3759.987573880" lastFinishedPulling="2025-11-24 14:49:13.307944743 +0000 UTC m=+3768.595702036" observedRunningTime="2025-11-24 14:49:13.822295067 +0000 UTC m=+3769.110052350" watchObservedRunningTime="2025-11-24 14:49:13.825000061 +0000 UTC m=+3769.112757354" Nov 24 14:49:16 crc kubenswrapper[4970]: I1124 14:49:16.102636 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_40d13307-a86f-453f-a3c8-05f936d03e73/memcached/0.log" Nov 24 14:49:22 crc kubenswrapper[4970]: I1124 14:49:22.471422 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:49:22 crc kubenswrapper[4970]: E1124 14:49:22.472814 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:49:23 crc kubenswrapper[4970]: I1124 14:49:23.364688 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:23 crc kubenswrapper[4970]: I1124 14:49:23.365028 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:23 crc kubenswrapper[4970]: I1124 14:49:23.426122 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:23 crc kubenswrapper[4970]: I1124 14:49:23.973387 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ghss2" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.067356 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ghss2"] Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.105618 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.105909 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bnws5" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="registry-server" containerID="cri-o://b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234" gracePeriod=2 Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.556233 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.664112 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbjlx\" (UniqueName: \"kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx\") pod \"962a33e9-70d8-463f-b70b-c0427b287b5a\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.664159 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content\") pod \"962a33e9-70d8-463f-b70b-c0427b287b5a\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.664281 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities\") pod \"962a33e9-70d8-463f-b70b-c0427b287b5a\" (UID: \"962a33e9-70d8-463f-b70b-c0427b287b5a\") " Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.665492 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities" (OuterVolumeSpecName: "utilities") pod "962a33e9-70d8-463f-b70b-c0427b287b5a" (UID: "962a33e9-70d8-463f-b70b-c0427b287b5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.672406 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx" (OuterVolumeSpecName: "kube-api-access-rbjlx") pod "962a33e9-70d8-463f-b70b-c0427b287b5a" (UID: "962a33e9-70d8-463f-b70b-c0427b287b5a"). InnerVolumeSpecName "kube-api-access-rbjlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.718805 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "962a33e9-70d8-463f-b70b-c0427b287b5a" (UID: "962a33e9-70d8-463f-b70b-c0427b287b5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.766754 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbjlx\" (UniqueName: \"kubernetes.io/projected/962a33e9-70d8-463f-b70b-c0427b287b5a-kube-api-access-rbjlx\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.766786 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.766796 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a33e9-70d8-463f-b70b-c0427b287b5a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.915315 4970 generic.go:334] "Generic (PLEG): container finished" podID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerID="b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234" exitCode=0 Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.915408 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bnws5" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.915398 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerDied","Data":"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234"} Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.915475 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bnws5" event={"ID":"962a33e9-70d8-463f-b70b-c0427b287b5a","Type":"ContainerDied","Data":"b367b218058f81ddbfe5d8430bc6f4bf807fcb4eba5c2a4bbc85c88790f31379"} Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.915499 4970 scope.go:117] "RemoveContainer" containerID="b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.951408 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.961215 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bnws5"] Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.962330 4970 scope.go:117] "RemoveContainer" containerID="d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd" Nov 24 14:49:24 crc kubenswrapper[4970]: I1124 14:49:24.994810 4970 scope.go:117] "RemoveContainer" containerID="1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.045494 4970 scope.go:117] "RemoveContainer" containerID="b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234" Nov 24 14:49:25 crc kubenswrapper[4970]: E1124 14:49:25.046251 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234\": container with ID starting with b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234 not found: ID does not exist" containerID="b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.046280 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234"} err="failed to get container status \"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234\": rpc error: code = NotFound desc = could not find container \"b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234\": container with ID starting with b9ce6b25756991802afb9111134a144caa6db77defb86d3abe6ddfc858a6e234 not found: ID does not exist" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.046301 4970 scope.go:117] "RemoveContainer" containerID="d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd" Nov 24 14:49:25 crc kubenswrapper[4970]: E1124 14:49:25.046536 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd\": container with ID starting with d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd not found: ID does not exist" containerID="d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.046594 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd"} err="failed to get container status \"d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd\": rpc error: code = NotFound desc = could not find container \"d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd\": container with ID starting with d3501cf7fc1f4b72654c08348c69a37629e862efe97ce4b16debb7acb1bb6edd not found: ID does not exist" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.046627 4970 scope.go:117] "RemoveContainer" containerID="1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a" Nov 24 14:49:25 crc kubenswrapper[4970]: E1124 14:49:25.047002 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a\": container with ID starting with 1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a not found: ID does not exist" containerID="1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.047055 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a"} err="failed to get container status \"1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a\": rpc error: code = NotFound desc = could not find container \"1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a\": container with ID starting with 1e141c1f23b1971d1cd5bee1156d0045f82a541c2a384eb878ce8a756bf7e02a not found: ID does not exist" Nov 24 14:49:25 crc kubenswrapper[4970]: I1124 14:49:25.483354 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" path="/var/lib/kubelet/pods/962a33e9-70d8-463f-b70b-c0427b287b5a/volumes" Nov 24 14:49:33 crc kubenswrapper[4970]: I1124 14:49:33.886614 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.205802 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.248673 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.291275 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.339141 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/util/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.418059 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/pull/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.453710 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1ca9b138781dcf125934bc878376abf75f877c2252ee8cf8f3500b7287xlnlr_373aeb29-00a5-4ee1-a007-fc9c113a99b6/extract/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.525447 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kdx5m_93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5/kube-rbac-proxy/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.666179 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7768f8c84f-kdx5m_93f2c52e-2ee5-4f46-af2d-af7cb9d3a2a5/manager/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.693881 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-r452c_8874383f-46c3-43b7-b1d1-56c9d14a0004/kube-rbac-proxy/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.780596 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6d8fd67bf7-r452c_8874383f-46c3-43b7-b1d1-56c9d14a0004/manager/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.860465 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-ff65l_799af0bf-24f0-4546-960b-2e8e72083fb8/kube-rbac-proxy/0.log" Nov 24 14:49:34 crc kubenswrapper[4970]: I1124 14:49:34.936536 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-56dfb6b67f-ff65l_799af0bf-24f0-4546-960b-2e8e72083fb8/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.077044 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qclm2_a5f4428d-ecd0-4c71-b9da-de188497a191/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.162439 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-8667fbf6f6-qclm2_a5f4428d-ecd0-4c71-b9da-de188497a191/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.172425 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-v67fx_9ca1ba14-aafe-429b-be3d-87dadeb73b54/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.340326 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-bf4c6585d-v67fx_9ca1ba14-aafe-429b-be3d-87dadeb73b54/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.356722 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-crqqj_4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.401727 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d86b44686-crqqj_4c47288e-7c2a-4c44-ba1d-f12fe6fecbdd/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.476176 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:49:35 crc kubenswrapper[4970]: E1124 14:49:35.476468 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.527996 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-9lhqx_d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.646808 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-769d9c7585-9lhqx_d3c5b8af-2db1-4b4c-8cd3-da2c2ff7fff0/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.755769 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-sx2v5_f01cc05b-92d6-4144-ad87-5fdec63cc95c/manager/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.758103 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5c75d7c94b-sx2v5_f01cc05b-92d6-4144-ad87-5fdec63cc95c/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.843365 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-9mnkp_20a43498-16e1-46d8-9487-b87ff9cad8f8/kube-rbac-proxy/0.log" Nov 24 14:49:35 crc kubenswrapper[4970]: I1124 14:49:35.991429 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7879fb76fd-9mnkp_20a43498-16e1-46d8-9487-b87ff9cad8f8/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.050468 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-h9j24_73f36b2f-3ac8-4d5c-859b-1e376b9a5194/kube-rbac-proxy/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.059045 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7bb88cb858-h9j24_73f36b2f-3ac8-4d5c-859b-1e376b9a5194/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.198320 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-wfsc5_b070979e-cf9f-4793-a50a-f0e67a48a58f/kube-rbac-proxy/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.256042 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6f8c5b86cb-wfsc5_b070979e-cf9f-4793-a50a-f0e67a48a58f/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.410197 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-zkcwk_028f94a9-1db9-46e8-93f1-106d959ebe9b/kube-rbac-proxy/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.524341 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-66b7d6f598-zkcwk_028f94a9-1db9-46e8-93f1-106d959ebe9b/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.535545 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-vs899_a44bf58b-6eb3-4bdc-ad1c-41aa7623666b/kube-rbac-proxy/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.712552 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-86d796d84d-vs899_a44bf58b-6eb3-4bdc-ad1c-41aa7623666b/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.794817 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-g8kmd_d94c9e13-b160-4f65-a97c-9bfdd4799eb0/manager/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.827615 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6fdc856c5d-g8kmd_d94c9e13-b160-4f65-a97c-9bfdd4799eb0/kube-rbac-proxy/0.log" Nov 24 14:49:36 crc kubenswrapper[4970]: I1124 14:49:36.990406 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp_d17a257a-f687-4d53-9b81-5fa430e24d70/kube-rbac-proxy/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.027749 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-79d88dcd44ct6wp_d17a257a-f687-4d53-9b81-5fa430e24d70/manager/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.102745 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-bcbxx_6ed5751d-33ce-4775-a0c7-cb951a6fbdcb/kube-rbac-proxy/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.353076 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-pb4vh_f268a9ff-bab7-41a4-9d10-d0494d511f3a/kube-rbac-proxy/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.428330 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-8486c7f98b-pb4vh_f268a9ff-bab7-41a4-9d10-d0494d511f3a/operator/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.685403 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wrstd_2bca4967-78c0-444a-bd32-e12f72f64629/registry-server/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.779891 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-hdflg_1af43cf6-5835-47eb-a315-c16bf5758a9e/kube-rbac-proxy/0.log" Nov 24 14:49:37 crc kubenswrapper[4970]: I1124 14:49:37.852461 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5bdf4f7f7f-hdflg_1af43cf6-5835-47eb-a315-c16bf5758a9e/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.059949 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-t4pqg_90e1030d-d9d0-42bf-b35b-a2f19ee0e6de/kube-rbac-proxy/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.069086 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-6dc664666c-t4pqg_90e1030d-d9d0-42bf-b35b-a2f19ee0e6de/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.204323 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-488cd_65008609-c137-4dee-b406-159814902ff4/operator/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.220469 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6cb9dc54f8-bcbxx_6ed5751d-33ce-4775-a0c7-cb951a6fbdcb/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.270791 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-5sp97_0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.280397 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-799cb6ffd6-5sp97_0a90d4cf-ecb1-4c0d-ab4d-b59d9a81cd20/kube-rbac-proxy/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.588432 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-k5b5j_289cc484-057a-4107-937d-6c532c21b52a/kube-rbac-proxy/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.633771 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7798859c74-k5b5j_289cc484-057a-4107-937d-6c532c21b52a/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.681557 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-wfmzk_5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7/kube-rbac-proxy/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.709085 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-8464cf66df-wfmzk_5f0f2e45-b8ee-4206-b23a-13f69eb3c5b7/manager/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.790290 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-m9m2l_9847f51b-ce73-493a-ac47-b93341cebf73/kube-rbac-proxy/0.log" Nov 24 14:49:38 crc kubenswrapper[4970]: I1124 14:49:38.815250 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7cd4fb6f79-m9m2l_9847f51b-ce73-493a-ac47-b93341cebf73/manager/0.log" Nov 24 14:49:47 crc kubenswrapper[4970]: I1124 14:49:47.470910 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:49:47 crc kubenswrapper[4970]: E1124 14:49:47.471900 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:49:56 crc kubenswrapper[4970]: I1124 14:49:56.046273 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xl78j_4b2d5731-75aa-457d-8afb-75815fd5ca93/control-plane-machine-set-operator/0.log" Nov 24 14:49:56 crc kubenswrapper[4970]: I1124 14:49:56.214140 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5jd8l_3999597f-f349-4448-b4f4-45fa925f3342/kube-rbac-proxy/0.log" Nov 24 14:49:56 crc kubenswrapper[4970]: I1124 14:49:56.220842 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5jd8l_3999597f-f349-4448-b4f4-45fa925f3342/machine-api-operator/0.log" Nov 24 14:49:59 crc kubenswrapper[4970]: I1124 14:49:59.471106 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:49:59 crc kubenswrapper[4970]: E1124 14:49:59.472150 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:50:10 crc kubenswrapper[4970]: I1124 14:50:10.652313 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-cjd48_45373789-c5a3-47eb-b2fe-7b4ce5236a33/cert-manager-controller/0.log" Nov 24 14:50:11 crc kubenswrapper[4970]: I1124 14:50:11.003879 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-22n4h_5da63529-7e53-4ad3-8d4b-0bc32d616af8/cert-manager-cainjector/0.log" Nov 24 14:50:11 crc kubenswrapper[4970]: I1124 14:50:11.061369 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-bz6q5_fe636fdb-0614-4cbc-a621-086b3c591926/cert-manager-webhook/0.log" Nov 24 14:50:14 crc kubenswrapper[4970]: I1124 14:50:14.471146 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:50:14 crc kubenswrapper[4970]: E1124 14:50:14.472242 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.215315 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-nwkjr_ebda0f09-edf4-40b6-a13d-f04cf3ddd510/nmstate-console-plugin/0.log" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.344616 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fpsdv_2e17e3dd-4f3c-4035-9c8f-83a37b7a2c48/nmstate-handler/0.log" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.441260 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8nmq4_62a1e00f-7e1a-4e97-a311-5628f78a7798/kube-rbac-proxy/0.log" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.447105 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8nmq4_62a1e00f-7e1a-4e97-a311-5628f78a7798/nmstate-metrics/0.log" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.668021 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-fdjzs_63629bdf-0eb6-46d8-a10b-08cc587486a1/nmstate-operator/0.log" Nov 24 14:50:25 crc kubenswrapper[4970]: I1124 14:50:25.684261 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-bjrv5_892ed300-f0bb-49aa-80e3-b7feb492606c/nmstate-webhook/0.log" Nov 24 14:50:28 crc kubenswrapper[4970]: I1124 14:50:28.471068 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:50:28 crc kubenswrapper[4970]: E1124 14:50:28.471552 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.562192 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-s8fl4_48658fb0-3b0d-4505-9e99-68d21fa216e8/kube-rbac-proxy/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.615230 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-s8fl4_48658fb0-3b0d-4505-9e99-68d21fa216e8/controller/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.726327 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.918296 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.936434 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.938978 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:50:40 crc kubenswrapper[4970]: I1124 14:50:40.946522 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.108871 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.108981 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.114157 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.117903 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.289430 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-frr-files/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.300082 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-metrics/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.300293 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/controller/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.307042 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/cp-reloader/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.466861 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/frr-metrics/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.519756 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/kube-rbac-proxy/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.525991 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/kube-rbac-proxy-frr/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.643604 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/reloader/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.748893 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-c767v_fdf01015-8aee-4c8d-bdd2-0a72e6387d32/frr-k8s-webhook-server/0.log" Nov 24 14:50:41 crc kubenswrapper[4970]: I1124 14:50:41.976308 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56999fcf9c-k9nw7_39678371-0ddb-40cb-8b7d-375d70868b75/manager/0.log" Nov 24 14:50:42 crc kubenswrapper[4970]: I1124 14:50:42.152248 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56df5758cc-gjrmn_db2d684a-3178-4375-b857-fc4ab433a328/webhook-server/0.log" Nov 24 14:50:42 crc kubenswrapper[4970]: I1124 14:50:42.220546 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qn5c2_2bc53266-821b-4f2b-af20-c290fff7e4a9/kube-rbac-proxy/0.log" Nov 24 14:50:42 crc kubenswrapper[4970]: I1124 14:50:42.716735 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qn5c2_2bc53266-821b-4f2b-af20-c290fff7e4a9/speaker/0.log" Nov 24 14:50:42 crc kubenswrapper[4970]: I1124 14:50:42.816802 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vgst9_9bd50032-57ea-42cc-9c4c-0a4213b02636/frr/0.log" Nov 24 14:50:43 crc kubenswrapper[4970]: I1124 14:50:43.470196 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:50:43 crc kubenswrapper[4970]: E1124 14:50:43.470450 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:50:55 crc kubenswrapper[4970]: I1124 14:50:55.478873 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:50:55 crc kubenswrapper[4970]: E1124 14:50:55.480507 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.107541 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.260054 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.300641 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.303729 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.442635 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/util/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.446264 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/pull/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.500828 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772erdspx_a21ef8b8-372b-45fb-93e3-d0318862eab0/extract/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.614413 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.742949 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.761065 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.761509 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:50:56 crc kubenswrapper[4970]: I1124 14:50:56.973774 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-content/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.000044 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/extract-utilities/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.175034 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-utilities/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.420838 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-content/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.433544 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-utilities/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.441910 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-content/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.644369 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-content/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.668014 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-w8gm8_745f7b30-991a-4df2-bdae-d7438d628a7f/registry-server/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.725804 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/extract-utilities/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.862463 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ghss2_8d0b0fbd-8b3b-4faa-90f8-02a413146f8a/registry-server/0.log" Nov 24 14:50:57 crc kubenswrapper[4970]: I1124 14:50:57.903507 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.052550 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.075750 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.076392 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.229860 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/pull/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.242523 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/util/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.269945 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6px6fc_a3958bff-b868-47ce-bd5d-7adb85d9dab0/extract/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.429049 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8njc2_9776cf3e-bec6-4eca-874b-37e8dc5a755e/marketplace-operator/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.449190 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.633651 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.646867 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.686133 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.878026 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-utilities/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.907570 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/extract-content/0.log" Nov 24 14:50:58 crc kubenswrapper[4970]: I1124 14:50:58.971598 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2lldf_17f79817-a5af-4690-8524-716178cec162/registry-server/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.078127 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.271760 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.282775 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.299427 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.495651 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-utilities/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.508858 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/extract-content/0.log" Nov 24 14:50:59 crc kubenswrapper[4970]: I1124 14:50:59.852292 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-ggqrv_002977b6-e9fd-4de3-8753-96a4d041c7fd/registry-server/0.log" Nov 24 14:51:09 crc kubenswrapper[4970]: I1124 14:51:09.471050 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:51:09 crc kubenswrapper[4970]: E1124 14:51:09.471693 4970 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9jtl9_openshift-machine-config-operator(5dd7b91d-1aca-41aa-b8b4-ab97723e8074)\"" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" Nov 24 14:51:24 crc kubenswrapper[4970]: I1124 14:51:24.471232 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:51:25 crc kubenswrapper[4970]: I1124 14:51:25.077097 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"ce571cc7ba4a9d7201e8a3026e3764b340dc01ba930640be9c7333520981c1b3"} Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.256565 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:20 crc kubenswrapper[4970]: E1124 14:52:20.258177 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="extract-utilities" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.258210 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="extract-utilities" Nov 24 14:52:20 crc kubenswrapper[4970]: E1124 14:52:20.258295 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="extract-content" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.258314 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="extract-content" Nov 24 14:52:20 crc kubenswrapper[4970]: E1124 14:52:20.258349 4970 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="registry-server" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.258366 4970 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="registry-server" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.258849 4970 memory_manager.go:354] "RemoveStaleState removing state" podUID="962a33e9-70d8-463f-b70b-c0427b287b5a" containerName="registry-server" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.262252 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.286312 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.319750 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6495\" (UniqueName: \"kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.319806 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.319964 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.421952 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6495\" (UniqueName: \"kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.422373 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.422634 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.422969 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.423172 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.452822 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6495\" (UniqueName: \"kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495\") pod \"certified-operators-9cm9x\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:20 crc kubenswrapper[4970]: I1124 14:52:20.601048 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:21 crc kubenswrapper[4970]: I1124 14:52:21.209740 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:21 crc kubenswrapper[4970]: I1124 14:52:21.751218 4970 generic.go:334] "Generic (PLEG): container finished" podID="fe159dcb-b8e7-45ad-9b36-106d1002210f" containerID="a16f1df46182bf67c2598ac814acf4e0342a23c97f2279bb75c4f84b0215d0d0" exitCode=0 Nov 24 14:52:21 crc kubenswrapper[4970]: I1124 14:52:21.751287 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerDied","Data":"a16f1df46182bf67c2598ac814acf4e0342a23c97f2279bb75c4f84b0215d0d0"} Nov 24 14:52:21 crc kubenswrapper[4970]: I1124 14:52:21.751331 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerStarted","Data":"f34acef752969f9c35c0ed50e6999197ae7d65a2bcb8c32487a2a5244b5f151f"} Nov 24 14:52:22 crc kubenswrapper[4970]: I1124 14:52:22.765456 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerStarted","Data":"2c003da4eb7b2d0d90b037de8d4d552f684af1d1809e8e873005b001b1844ead"} Nov 24 14:52:23 crc kubenswrapper[4970]: I1124 14:52:23.779381 4970 generic.go:334] "Generic (PLEG): container finished" podID="fe159dcb-b8e7-45ad-9b36-106d1002210f" containerID="2c003da4eb7b2d0d90b037de8d4d552f684af1d1809e8e873005b001b1844ead" exitCode=0 Nov 24 14:52:23 crc kubenswrapper[4970]: I1124 14:52:23.779431 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerDied","Data":"2c003da4eb7b2d0d90b037de8d4d552f684af1d1809e8e873005b001b1844ead"} Nov 24 14:52:24 crc kubenswrapper[4970]: I1124 14:52:24.795924 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerStarted","Data":"a50bec60624388ad99aaf17af28cf9d471e776a092af111bf53fd80de3c6153f"} Nov 24 14:52:24 crc kubenswrapper[4970]: I1124 14:52:24.820362 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9cm9x" podStartSLOduration=2.371101276 podStartE2EDuration="4.820341692s" podCreationTimestamp="2025-11-24 14:52:20 +0000 UTC" firstStartedPulling="2025-11-24 14:52:21.753546075 +0000 UTC m=+3957.041303408" lastFinishedPulling="2025-11-24 14:52:24.202786491 +0000 UTC m=+3959.490543824" observedRunningTime="2025-11-24 14:52:24.813353412 +0000 UTC m=+3960.101110715" watchObservedRunningTime="2025-11-24 14:52:24.820341692 +0000 UTC m=+3960.108098985" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.238411 4970 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.241794 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.294814 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.340945 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw22t\" (UniqueName: \"kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.341020 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.341143 4970 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.442862 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw22t\" (UniqueName: \"kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.442927 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.443028 4970 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.443644 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.443698 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.481058 4970 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw22t\" (UniqueName: \"kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t\") pod \"redhat-marketplace-msh7h\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:29 crc kubenswrapper[4970]: I1124 14:52:29.606132 4970 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.061280 4970 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.602339 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.602820 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.689854 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.874978 4970 generic.go:334] "Generic (PLEG): container finished" podID="edad337a-5f5a-490d-8496-3fa6357f00a0" containerID="7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4" exitCode=0 Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.875095 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerDied","Data":"7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4"} Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.875153 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerStarted","Data":"fce4a9151c7b27e48f5ab4ff7d79e16b18d62d36554b77bf568a977400e4982c"} Nov 24 14:52:30 crc kubenswrapper[4970]: I1124 14:52:30.966703 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:31 crc kubenswrapper[4970]: I1124 14:52:31.887219 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerStarted","Data":"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3"} Nov 24 14:52:32 crc kubenswrapper[4970]: I1124 14:52:32.905690 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerDied","Data":"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3"} Nov 24 14:52:32 crc kubenswrapper[4970]: I1124 14:52:32.905487 4970 generic.go:334] "Generic (PLEG): container finished" podID="edad337a-5f5a-490d-8496-3fa6357f00a0" containerID="5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3" exitCode=0 Nov 24 14:52:33 crc kubenswrapper[4970]: I1124 14:52:33.019661 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:33 crc kubenswrapper[4970]: I1124 14:52:33.020978 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9cm9x" podUID="fe159dcb-b8e7-45ad-9b36-106d1002210f" containerName="registry-server" containerID="cri-o://a50bec60624388ad99aaf17af28cf9d471e776a092af111bf53fd80de3c6153f" gracePeriod=2 Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:33.931685 4970 generic.go:334] "Generic (PLEG): container finished" podID="fe159dcb-b8e7-45ad-9b36-106d1002210f" containerID="a50bec60624388ad99aaf17af28cf9d471e776a092af111bf53fd80de3c6153f" exitCode=0 Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:33.932108 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerDied","Data":"a50bec60624388ad99aaf17af28cf9d471e776a092af111bf53fd80de3c6153f"} Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.091654 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.140115 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content\") pod \"fe159dcb-b8e7-45ad-9b36-106d1002210f\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.140299 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities\") pod \"fe159dcb-b8e7-45ad-9b36-106d1002210f\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.140320 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6495\" (UniqueName: \"kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495\") pod \"fe159dcb-b8e7-45ad-9b36-106d1002210f\" (UID: \"fe159dcb-b8e7-45ad-9b36-106d1002210f\") " Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.142028 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities" (OuterVolumeSpecName: "utilities") pod "fe159dcb-b8e7-45ad-9b36-106d1002210f" (UID: "fe159dcb-b8e7-45ad-9b36-106d1002210f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.150760 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495" (OuterVolumeSpecName: "kube-api-access-k6495") pod "fe159dcb-b8e7-45ad-9b36-106d1002210f" (UID: "fe159dcb-b8e7-45ad-9b36-106d1002210f"). InnerVolumeSpecName "kube-api-access-k6495". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.193427 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe159dcb-b8e7-45ad-9b36-106d1002210f" (UID: "fe159dcb-b8e7-45ad-9b36-106d1002210f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.242986 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.243007 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe159dcb-b8e7-45ad-9b36-106d1002210f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.243016 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6495\" (UniqueName: \"kubernetes.io/projected/fe159dcb-b8e7-45ad-9b36-106d1002210f-kube-api-access-k6495\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.950855 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cm9x" event={"ID":"fe159dcb-b8e7-45ad-9b36-106d1002210f","Type":"ContainerDied","Data":"f34acef752969f9c35c0ed50e6999197ae7d65a2bcb8c32487a2a5244b5f151f"} Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.951767 4970 scope.go:117] "RemoveContainer" containerID="a50bec60624388ad99aaf17af28cf9d471e776a092af111bf53fd80de3c6153f" Nov 24 14:52:34 crc kubenswrapper[4970]: I1124 14:52:34.951155 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cm9x" Nov 24 14:52:35 crc kubenswrapper[4970]: I1124 14:52:35.003628 4970 scope.go:117] "RemoveContainer" containerID="2c003da4eb7b2d0d90b037de8d4d552f684af1d1809e8e873005b001b1844ead" Nov 24 14:52:35 crc kubenswrapper[4970]: I1124 14:52:35.019096 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:35 crc kubenswrapper[4970]: I1124 14:52:35.033867 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9cm9x"] Nov 24 14:52:35 crc kubenswrapper[4970]: I1124 14:52:35.041405 4970 scope.go:117] "RemoveContainer" containerID="a16f1df46182bf67c2598ac814acf4e0342a23c97f2279bb75c4f84b0215d0d0" Nov 24 14:52:35 crc kubenswrapper[4970]: I1124 14:52:35.488623 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe159dcb-b8e7-45ad-9b36-106d1002210f" path="/var/lib/kubelet/pods/fe159dcb-b8e7-45ad-9b36-106d1002210f/volumes" Nov 24 14:52:36 crc kubenswrapper[4970]: I1124 14:52:36.975904 4970 generic.go:334] "Generic (PLEG): container finished" podID="f26a4374-de0f-42db-abdc-3cbec3f634a0" containerID="41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859" exitCode=0 Nov 24 14:52:36 crc kubenswrapper[4970]: I1124 14:52:36.976009 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4gvrb/must-gather-9mwph" event={"ID":"f26a4374-de0f-42db-abdc-3cbec3f634a0","Type":"ContainerDied","Data":"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859"} Nov 24 14:52:36 crc kubenswrapper[4970]: I1124 14:52:36.976831 4970 scope.go:117] "RemoveContainer" containerID="41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859" Nov 24 14:52:37 crc kubenswrapper[4970]: I1124 14:52:37.575627 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gvrb_must-gather-9mwph_f26a4374-de0f-42db-abdc-3cbec3f634a0/gather/0.log" Nov 24 14:52:37 crc kubenswrapper[4970]: I1124 14:52:37.989021 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerStarted","Data":"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936"} Nov 24 14:52:38 crc kubenswrapper[4970]: I1124 14:52:38.014224 4970 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-msh7h" podStartSLOduration=2.512131669 podStartE2EDuration="9.014204938s" podCreationTimestamp="2025-11-24 14:52:29 +0000 UTC" firstStartedPulling="2025-11-24 14:52:30.877906105 +0000 UTC m=+3966.165663428" lastFinishedPulling="2025-11-24 14:52:37.379979394 +0000 UTC m=+3972.667736697" observedRunningTime="2025-11-24 14:52:38.012133551 +0000 UTC m=+3973.299890884" watchObservedRunningTime="2025-11-24 14:52:38.014204938 +0000 UTC m=+3973.301962231" Nov 24 14:52:39 crc kubenswrapper[4970]: I1124 14:52:39.606301 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:39 crc kubenswrapper[4970]: I1124 14:52:39.606713 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:39 crc kubenswrapper[4970]: I1124 14:52:39.689025 4970 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.258852 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4gvrb/must-gather-9mwph"] Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.259501 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4gvrb/must-gather-9mwph" podUID="f26a4374-de0f-42db-abdc-3cbec3f634a0" containerName="copy" containerID="cri-o://97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713" gracePeriod=2 Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.269297 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4gvrb/must-gather-9mwph"] Nov 24 14:52:48 crc kubenswrapper[4970]: E1124 14:52:48.459961 4970 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf26a4374_de0f_42db_abdc_3cbec3f634a0.slice/crio-conmon-97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713.scope\": RecentStats: unable to find data in memory cache]" Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.693677 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gvrb_must-gather-9mwph_f26a4374-de0f-42db-abdc-3cbec3f634a0/copy/0.log" Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.695737 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.897773 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output\") pod \"f26a4374-de0f-42db-abdc-3cbec3f634a0\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.898275 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtpj2\" (UniqueName: \"kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2\") pod \"f26a4374-de0f-42db-abdc-3cbec3f634a0\" (UID: \"f26a4374-de0f-42db-abdc-3cbec3f634a0\") " Nov 24 14:52:48 crc kubenswrapper[4970]: I1124 14:52:48.912190 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2" (OuterVolumeSpecName: "kube-api-access-gtpj2") pod "f26a4374-de0f-42db-abdc-3cbec3f634a0" (UID: "f26a4374-de0f-42db-abdc-3cbec3f634a0"). InnerVolumeSpecName "kube-api-access-gtpj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.000343 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtpj2\" (UniqueName: \"kubernetes.io/projected/f26a4374-de0f-42db-abdc-3cbec3f634a0-kube-api-access-gtpj2\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.064296 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "f26a4374-de0f-42db-abdc-3cbec3f634a0" (UID: "f26a4374-de0f-42db-abdc-3cbec3f634a0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.101757 4970 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/f26a4374-de0f-42db-abdc-3cbec3f634a0-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.115803 4970 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4gvrb_must-gather-9mwph_f26a4374-de0f-42db-abdc-3cbec3f634a0/copy/0.log" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.116202 4970 generic.go:334] "Generic (PLEG): container finished" podID="f26a4374-de0f-42db-abdc-3cbec3f634a0" containerID="97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713" exitCode=143 Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.116253 4970 scope.go:117] "RemoveContainer" containerID="97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.116258 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4gvrb/must-gather-9mwph" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.134590 4970 scope.go:117] "RemoveContainer" containerID="41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.193348 4970 scope.go:117] "RemoveContainer" containerID="97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713" Nov 24 14:52:49 crc kubenswrapper[4970]: E1124 14:52:49.193799 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713\": container with ID starting with 97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713 not found: ID does not exist" containerID="97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.193829 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713"} err="failed to get container status \"97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713\": rpc error: code = NotFound desc = could not find container \"97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713\": container with ID starting with 97d720a53884735216cd8a722734577990f52ca79ac60412579edd711f529713 not found: ID does not exist" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.193848 4970 scope.go:117] "RemoveContainer" containerID="41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859" Nov 24 14:52:49 crc kubenswrapper[4970]: E1124 14:52:49.194236 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859\": container with ID starting with 41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859 not found: ID does not exist" containerID="41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.194260 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859"} err="failed to get container status \"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859\": rpc error: code = NotFound desc = could not find container \"41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859\": container with ID starting with 41af97ef26969f8fb4fbee3ff2bc782ac8f1430b91f8c29ff52ccde62de59859 not found: ID does not exist" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.486359 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f26a4374-de0f-42db-abdc-3cbec3f634a0" path="/var/lib/kubelet/pods/f26a4374-de0f-42db-abdc-3cbec3f634a0/volumes" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.687410 4970 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:49 crc kubenswrapper[4970]: I1124 14:52:49.745345 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.127763 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-msh7h" podUID="edad337a-5f5a-490d-8496-3fa6357f00a0" containerName="registry-server" containerID="cri-o://3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936" gracePeriod=2 Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.664772 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.831741 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content\") pod \"edad337a-5f5a-490d-8496-3fa6357f00a0\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.831847 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities\") pod \"edad337a-5f5a-490d-8496-3fa6357f00a0\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.831897 4970 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw22t\" (UniqueName: \"kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t\") pod \"edad337a-5f5a-490d-8496-3fa6357f00a0\" (UID: \"edad337a-5f5a-490d-8496-3fa6357f00a0\") " Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.833257 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities" (OuterVolumeSpecName: "utilities") pod "edad337a-5f5a-490d-8496-3fa6357f00a0" (UID: "edad337a-5f5a-490d-8496-3fa6357f00a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.840884 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t" (OuterVolumeSpecName: "kube-api-access-hw22t") pod "edad337a-5f5a-490d-8496-3fa6357f00a0" (UID: "edad337a-5f5a-490d-8496-3fa6357f00a0"). InnerVolumeSpecName "kube-api-access-hw22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.879445 4970 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "edad337a-5f5a-490d-8496-3fa6357f00a0" (UID: "edad337a-5f5a-490d-8496-3fa6357f00a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.934797 4970 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.934831 4970 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/edad337a-5f5a-490d-8496-3fa6357f00a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:50 crc kubenswrapper[4970]: I1124 14:52:50.934846 4970 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw22t\" (UniqueName: \"kubernetes.io/projected/edad337a-5f5a-490d-8496-3fa6357f00a0-kube-api-access-hw22t\") on node \"crc\" DevicePath \"\"" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.140773 4970 generic.go:334] "Generic (PLEG): container finished" podID="edad337a-5f5a-490d-8496-3fa6357f00a0" containerID="3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936" exitCode=0 Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.140838 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerDied","Data":"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936"} Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.140964 4970 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-msh7h" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.141433 4970 scope.go:117] "RemoveContainer" containerID="3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.160660 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-msh7h" event={"ID":"edad337a-5f5a-490d-8496-3fa6357f00a0","Type":"ContainerDied","Data":"fce4a9151c7b27e48f5ab4ff7d79e16b18d62d36554b77bf568a977400e4982c"} Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.177645 4970 scope.go:117] "RemoveContainer" containerID="5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.191952 4970 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.218110 4970 scope.go:117] "RemoveContainer" containerID="7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.223714 4970 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-msh7h"] Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.276571 4970 scope.go:117] "RemoveContainer" containerID="3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936" Nov 24 14:52:51 crc kubenswrapper[4970]: E1124 14:52:51.277377 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936\": container with ID starting with 3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936 not found: ID does not exist" containerID="3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.277476 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936"} err="failed to get container status \"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936\": rpc error: code = NotFound desc = could not find container \"3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936\": container with ID starting with 3eae80db4c5525c3aa4bfe097f6a1cf292e11b8ebcfb1ba493e9ad058d265936 not found: ID does not exist" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.277559 4970 scope.go:117] "RemoveContainer" containerID="5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3" Nov 24 14:52:51 crc kubenswrapper[4970]: E1124 14:52:51.277988 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3\": container with ID starting with 5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3 not found: ID does not exist" containerID="5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.278028 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3"} err="failed to get container status \"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3\": rpc error: code = NotFound desc = could not find container \"5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3\": container with ID starting with 5ac3878e27fe8b528d49e5a3c254c9e2b3f18b402806a98bd86e647185fb52b3 not found: ID does not exist" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.278055 4970 scope.go:117] "RemoveContainer" containerID="7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4" Nov 24 14:52:51 crc kubenswrapper[4970]: E1124 14:52:51.278427 4970 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4\": container with ID starting with 7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4 not found: ID does not exist" containerID="7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.278493 4970 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4"} err="failed to get container status \"7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4\": rpc error: code = NotFound desc = could not find container \"7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4\": container with ID starting with 7dd9f8726fb3b8c5045e936af35546b47d9334ce7ff607856401e080ac241eb4 not found: ID does not exist" Nov 24 14:52:51 crc kubenswrapper[4970]: I1124 14:52:51.483315 4970 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edad337a-5f5a-490d-8496-3fa6357f00a0" path="/var/lib/kubelet/pods/edad337a-5f5a-490d-8496-3fa6357f00a0/volumes" Nov 24 14:53:41 crc kubenswrapper[4970]: I1124 14:53:41.204679 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:53:41 crc kubenswrapper[4970]: I1124 14:53:41.205293 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:54:11 crc kubenswrapper[4970]: I1124 14:54:11.204697 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:54:11 crc kubenswrapper[4970]: I1124 14:54:11.205188 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.204026 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.204851 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.204925 4970 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.206154 4970 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ce571cc7ba4a9d7201e8a3026e3764b340dc01ba930640be9c7333520981c1b3"} pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.206278 4970 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" containerID="cri-o://ce571cc7ba4a9d7201e8a3026e3764b340dc01ba930640be9c7333520981c1b3" gracePeriod=600 Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.438425 4970 generic.go:334] "Generic (PLEG): container finished" podID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerID="ce571cc7ba4a9d7201e8a3026e3764b340dc01ba930640be9c7333520981c1b3" exitCode=0 Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.438756 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerDied","Data":"ce571cc7ba4a9d7201e8a3026e3764b340dc01ba930640be9c7333520981c1b3"} Nov 24 14:54:41 crc kubenswrapper[4970]: I1124 14:54:41.438788 4970 scope.go:117] "RemoveContainer" containerID="79b5110f2366f7a844c73564a0be2d9e106400d61593135110b40cf2cb0b4143" Nov 24 14:54:42 crc kubenswrapper[4970]: I1124 14:54:42.467771 4970 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" event={"ID":"5dd7b91d-1aca-41aa-b8b4-ab97723e8074","Type":"ContainerStarted","Data":"6ccedccfb7189252ef7325734442c95d55d167bcdff1a2b73425920471b4ebbd"} Nov 24 14:54:48 crc kubenswrapper[4970]: I1124 14:54:48.281259 4970 scope.go:117] "RemoveContainer" containerID="93b8397f6128352afb8f3fb239c1f6362845cef937ca55a6502ce898d308a0dc" Nov 24 14:56:41 crc kubenswrapper[4970]: I1124 14:56:41.204669 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:56:41 crc kubenswrapper[4970]: I1124 14:56:41.205265 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 14:57:11 crc kubenswrapper[4970]: I1124 14:57:11.204673 4970 patch_prober.go:28] interesting pod/machine-config-daemon-9jtl9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 14:57:11 crc kubenswrapper[4970]: I1124 14:57:11.205337 4970 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9jtl9" podUID="5dd7b91d-1aca-41aa-b8b4-ab97723e8074" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111071122024434 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111071123017352 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111060345016501 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111060345015451 5ustar corecore